diff --git a/.gitattributes b/.gitattributes index 687381f0fdc24de0bd6a6ed0bfe5849e28bc310c..1f2149ebfb5ad85679c366001d6a901ffbdec46b 100644 --- a/.gitattributes +++ b/.gitattributes @@ -150,3 +150,93 @@ HVU_QA/t5-viet-qg-finetuned/checkpoint-24000/spiece.model filter=lfs diff=lfs me HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/model.safetensors filter=lfs diff=lfs merge=lfs -text HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/optimizer.pt filter=lfs diff=lfs merge=lfs -text HVU_QA/t5-viet-qg-finetuned/checkpoint-24500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/spiece.model filter=lfs diff=lfs merge=lfs -text diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e48203677494edcb4819b67547f5c6fd0a9e3d6f --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be7e97b0554b8185b2db8f4b5f58bfd9cfedc3f31e8c9a7a8beeefc6e63737c2 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..e6959a47c8155db3e7ae06a24fa3e129d131c424 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94c4ab75dc6ebca82847e075bb1d49b6ba6f92aa4f4705f458c53f1e4fa432af +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..1facb530557d859c595882ae0714beb80ece28d8 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..fc39d03918c64f811fe73e0d34d5b2015d6f251d Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..85336ba9caab942952ade7d31a966041b5d95a25 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/trainer_state.json @@ -0,0 +1,17534 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.965288235066991, + "eval_steps": 500, + "global_step": 25000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.5223947264e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8f946a952ca36bf449b0da4d3cb5ef041083ab86 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5acbf45084064d3418287de31ba031fc85e7bc787d5e3678b648e20eada40e7d +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..1961979d9ee4c40257611d3e810d7c45d1e87c1e --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c226a678f9e2195229ac4c02e5482987abd0370c969088d33415041242748167 +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..428a1c9b93c1f612a6e5de03682e6d073a4ecfa8 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..4458422d4d07722ae12427ef64a4674beb128722 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..b44aed1c2ee432a52e171bc6af30a98147c266fd --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/trainer_state.json @@ -0,0 +1,17884 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9845939997683308, + "eval_steps": 500, + "global_step": 25500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.552842620928e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-25500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..01bb46e50510a21dd188c311eda802c29dda3009 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e5f2f73964a674ee89420f4a15b0dbb251d0bbff3b0b36680f7abf42e033894d +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..56c4e31faa02bb2ff62c675b7ea68fd8262c9631 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:459510f1822d5caa1603d8760e5e37d786e5399dc510a8019449913db66dc693 +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..847565d92dd15de46d411251ba0749868bf5eff8 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..388f9413d994dc094438c6031d70296949a15501 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..b78b8d9bb7915556cf56d7596188cb3cd0d8cd05 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/trainer_state.json @@ -0,0 +1,18234 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0038997644696706, + "eval_steps": 500, + "global_step": 26000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.583290515456e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a0f024bba167aa396aa2d39e84c64796a9bea1b6 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f10ce85cd5d7b33162149da0aece7ecaacd09f95f1961c5453262b2dca443728 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..a3071630d9e0dca892a7da4da8f1a4a00db042ec --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:170bd22ae29f9937b6a9b2b7c3aa6d752e70efa445bd6d44b10d38e5d35c5b0b +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..adaff987c58fe8b9b8bcdb45542c250d5e4a37b8 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..d046897c9b0f6198f2af8b91106ace7fe351d16a Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..55bec2e208b3f6f4dbe9e0c5bfb1517b5aaef0a1 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/trainer_state.json @@ -0,0 +1,18584 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0232055291710105, + "eval_steps": 500, + "global_step": 26500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.613738409984e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2f3d586cc31f43b285b0bb2169533cecba7e6acd --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26db283046459b9bbbec060fee6b8e6fe193838ded75964d7b147130c3ca90b6 +size 292290560 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..3e5c5ee19877b40210996c3f9dd806c71e5b8b3c --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:166bf9712ab374ba28acf9ae0dcd301c9fa7a6b29b2b7992b5c21ad51c746af1 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c3d547146b0059cca9147d89362b056987a0f1ae Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..e5e4445c9a2d673cb557eb714354db4dcec1560c Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..69373152c1329a8b72944bf9d3bf407814ee6175 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/trainer_state.json @@ -0,0 +1,18934 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0425112938723502, + "eval_steps": 500, + "global_step": 27000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.644186304512e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0c436857901b7a24b173f90bc7616e46d63f4d77 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c7404c086c181cf347142444a7cf62496c878cb8dba1877c1b441277c65f6397 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..4a479adc545b73091832e9f666aefcf09dbacef8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ec150c7207f6b69c7f59a759b6e10d25ff9af447bc355636b36ec243ed49de5 +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..3cfaa509df3baa1f60c9591e628c64a309f33707 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..d469ca685876f54869a4c462606f4001018148eb Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..7c8f61840038e9ec656000ad9a3a8cb9755dead7 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/trainer_state.json @@ -0,0 +1,19284 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0618170585736901, + "eval_steps": 500, + "global_step": 27500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.67463419904e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..07c37760ea38ec76c4022e2eab79117445cf54fc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5be37576820cc1890d4b4095b5457ec6b0b3606663e17bccd1f6a07252689f25 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..e0c742b26fcf23385838bf59b188a33b39949281 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86511de074790b6181cfada2b7a0a7f053b9b3df4e546278b70f899f8ef62f38 +size 292552704 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..815c8d224174cc58137aa0165d1dbbfd384949ed Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..a54b6019a6555c46e9323a0e5c7ad56a7fe407aa Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..c9c6a1b50bcfed25831a034eb06c79e2b618cb40 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/trainer_state.json @@ -0,0 +1,19634 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0811228232750298, + "eval_steps": 500, + "global_step": 28000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.705082093568e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3c8a2048a549b3bd46d69da52653a481f95b092b --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25117a493ad2e10d17b8c24109652fa79ac0d8a652226dbb7f0ce1d4bdfb6ecc +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..b74ec2d4b893227a906f3ce9d641bf5954845970 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4819f80651afbbf8159fc7464c17612539fb4231c00d6c82203ffbf5e5083b92 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..ccad77c5debea75d7ef8fb41d2450889dca1e8d1 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..8806eabc4ae92b7aaa9ad4459122d073384c352c Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..b04a5fe92e2d0aa1f9e18b0ed71aa2723331e260 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/trainer_state.json @@ -0,0 +1,19984 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.1004285879763698, + "eval_steps": 500, + "global_step": 28500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.735529988096e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..be1975c1e18eba84d9d53f870f39a164c915c99f --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53c07322cb54eef4ee2761d496aa694016f1cd4323f5321d5c92e525c5fab84e +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..19389f8550a2f7e9e916e95b834ca5e6958cac30 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dff1d88feaa7937b57135cd34e66e724dce4ec1091b38cd27081b4531200c46b +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..dc0c893e15037e44dee9257538e9689d97717713 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..654a6d380491ee962df1eb501b3d73fed8b81c14 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..5ddaf72e254532dc287cc2cd5e8012f3a264abd0 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/trainer_state.json @@ -0,0 +1,20334 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.1197343526777095, + "eval_steps": 500, + "global_step": 29000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.765977882624e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cc7ad47d3b6eeac3aa7ef0e509f51fa5d1e6bf2d --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15123f197a6ffe7d4caa68fe5f424544a2603a0376ef0be4af35a23d3d19d068 +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..314aae9e634473d6df0d71f1f9399b6492f96c2c --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cfce9320e50ff48c9398b533bf56b434346a4cf774a95e2c4790fcf7789a7902 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..0ab44fb1e46b1992374ea25a45f8d5d05d0247dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..023c2684c5816851c28e7218f36481f029e119a5 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..f079fd356d3652f98c09f37b272db8f1a0f1b050 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/trainer_state.json @@ -0,0 +1,20684 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.1390401173790494, + "eval_steps": 500, + "global_step": 29500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.796425777152e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a45c708518ef76d2623686eece5ced951a6fcd7b --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0e6d8374f79de7713671bbaac634c962f8f6e2d38a11d8dda2aff795820e6b0 +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..ee2b1ce882a4cc57255aac51d01f017d7bb30b74 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c60a85305c64fc0c26b12beca372ce4e7e2da98942efc3c9f8759d450a4291cc +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..06aef96f997da4bdafb79bcf9a13359f5d0c7eab Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..04024603c666dc53630779eb176a6310fe8e53b7 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..5fe949b4589c2c3dc3010f0cd4d9557ad6df4619 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/trainer_state.json @@ -0,0 +1,21034 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.1583458820803891, + "eval_steps": 500, + "global_step": 30000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.82687367168e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..64e2f48e96b4b10c38b064ab5e78ba235b5b164c --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e13860353139711eaa0a02d6dc7cd9bc0591691974b39bdb537b369cdad1dacd +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..3bd7649095bc4e47c1642ae91f43cd7cdc4363b7 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9f1767d92c878f518a3d18b5b3d5e6c2730a6a703cc3a84874bd001a19a78926 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..fbd4f7b13b33ceeb1ee87895fbaca58c91384d67 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..cc8d447f3cdb9572d55f319c3c68e852fdcd9326 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..bbd9edf03916c0b324da8022bcb67bcf718fd0e7 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/trainer_state.json @@ -0,0 +1,21384 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.177651646781729, + "eval_steps": 500, + "global_step": 30500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.857321566208e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7ad305fb588a5fe57cee367953d050dddeeb448f --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf98be3d48213ae11d350a02dc984738adfa21074f50c837cb4cb1197f9863c2 +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..1c8e2c45dbd3b272ba7e218f127e7f08a757d6a2 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4af407162f108b99064abd20d5a16e01fe2aa317e454fe3562ab324fb8c0729 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..a4e9d034111fba4e294ecf03edc4f64d860c44a7 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..0ad2eee2951553de124e102ab37f5f85699408ac Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..d1363b08456c4dfad604b18db2bc161dcbd22ba2 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/trainer_state.json @@ -0,0 +1,21734 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.1969574114830688, + "eval_steps": 500, + "global_step": 31000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.887769460736e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..97633a557ec40864d9c6230146a380c82a264304 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dcf807af0df1d95d2f01d73bbf0fcddbb31d615792756a94965c6de72b9ad42b +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..0daefaf95d6d544a8165d29e844112e372d824ec --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67d632b4cdb224598e0427af4d2b2beca2f813de5f95461b81f27a53e04c3531 +size 292028416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c5980446d1e9becc6bcffb15fccfe64569fb4de7 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..a953e488614f08c79479045a46a71a8aff5fecf1 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..f4eb0e2eafc1e2f301b7598fd850bd177920feea --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/trainer_state.json @@ -0,0 +1,22084 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.2162631761844087, + "eval_steps": 500, + "global_step": 31500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.918217355264e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..141e0b6392b7799be93b975d1ae39b35e4f848bd --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43404e8c24e2696322e2472282796ae0fce7e04e4db778ebcc83a68823586ff0 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..de75c11488815fc74787606c1d25953169a1d947 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b62846d0e750572c2912df8e90adf021e468bc8b54d0828b8a1b2ce1ba3c0d96 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..231ab37a5c0e4530db1e508fa2b4251aab54e6d4 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..0afc0170107b0aaf9c2c8dc7f73583be6e49a720 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..93b4f894d6d204756cbe0742374ed96b712b2dbf --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/trainer_state.json @@ -0,0 +1,22434 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.2355689408857484, + "eval_steps": 500, + "global_step": 32000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.948665249792e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..837135b881c150f75c85d5d12075ca0848515b80 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e306ea49e70534ec8b5dcb1a52ed955e157767b6220718e3e182d64126e0c83 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..e7c38e4368dfb291c9f61edb726055af868a9e0e --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d697f585f051dbbf9b333d619ee4088cb85eac7a81df7cddc33a42e2ff6288d +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..f410699d5a837481e3a5649c5a969f4c6fdbec35 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..ccf8df01627e4848ee3a854ec885f9eedcd840a1 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..59acc498108e6ebc5ffaeb635ef2a9ad7685fd99 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/trainer_state.json @@ -0,0 +1,22784 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.2548747055870884, + "eval_steps": 500, + "global_step": 32500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.97911314432e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..78f7a546168a4a05c30223a5aabc17e16b807984 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e6398e60723f56dd4b36354cc9de6dddbf2b3dd2564e27af41d51b9dc76d03a +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..ecf88886edb7eb2ab60f27da568403032eccb685 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77af31ef5b6580922a7c5642e9afe001651eb232ae3705f7f9868b51693bc53b +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..972b666dd02f2b7ffe3af850cb3057bffa69c800 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..3c5be3da292027977c99cc0c26cf6c2d78de250c Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..d02fe2c3dcb7a426064e3a0ece7c064a6c1a3492 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/trainer_state.json @@ -0,0 +1,23134 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.274180470288428, + "eval_steps": 500, + "global_step": 33000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.009561038848e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2da31209b97a90c35b9674da16a21f1b5b82281d --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8311b0dd2d8df54c9a220e103bfd8d18a769f14fc9d3c17fabb0dee6501857e3 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..ea156bab87583a35842a1945e92ae6990679a41e --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a39a72bed8a331b59657b08361bddd16c00017fdc4ec67ef4c9cd9a80dc7e2f +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..b07e1a8d76813ee82151e6e0f70081b691f537f3 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..a218aba3526f906d61c8a253ad7f22a2176fff4f Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..487037a9b1a7f0f01cd2d6fb040cbf3ba41bc145 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/trainer_state.json @@ -0,0 +1,23484 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.293486234989768, + "eval_steps": 500, + "global_step": 33500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + }, + { + "epoch": 1.274566585582455, + "grad_norm": 0.5019121766090393, + "learning_rate": 0.0001150314683964632, + "loss": 0.1857, + "step": 33010 + }, + { + "epoch": 1.2749527008764816, + "grad_norm": 0.258952260017395, + "learning_rate": 0.0001150057273768614, + "loss": 0.11, + "step": 33020 + }, + { + "epoch": 1.2753388161705086, + "grad_norm": 0.8540348410606384, + "learning_rate": 0.00011497998635725962, + "loss": 0.1852, + "step": 33030 + }, + { + "epoch": 1.2757249314645354, + "grad_norm": 0.08689398318529129, + "learning_rate": 0.00011495424533765783, + "loss": 0.2225, + "step": 33040 + }, + { + "epoch": 1.2761110467585621, + "grad_norm": 0.42253822088241577, + "learning_rate": 0.00011492850431805605, + "loss": 0.0751, + "step": 33050 + }, + { + "epoch": 1.276497162052589, + "grad_norm": 1.2964017391204834, + "learning_rate": 0.00011490276329845426, + "loss": 0.2384, + "step": 33060 + }, + { + "epoch": 1.2768832773466157, + "grad_norm": 0.5337836146354675, + "learning_rate": 0.00011487702227885246, + "loss": 0.1415, + "step": 33070 + }, + { + "epoch": 1.2772693926406424, + "grad_norm": 2.7771682739257812, + "learning_rate": 0.00011485128125925069, + "loss": 0.378, + "step": 33080 + }, + { + "epoch": 1.2776555079346692, + "grad_norm": 1.5107232332229614, + "learning_rate": 0.00011482554023964889, + "loss": 0.2482, + "step": 33090 + }, + { + "epoch": 1.2780416232286962, + "grad_norm": 0.6885499358177185, + "learning_rate": 0.00011479979922004711, + "loss": 0.2122, + "step": 33100 + }, + { + "epoch": 1.278427738522723, + "grad_norm": 0.9016557335853577, + "learning_rate": 0.00011477405820044533, + "loss": 0.2841, + "step": 33110 + }, + { + "epoch": 1.2788138538167497, + "grad_norm": 1.9532525539398193, + "learning_rate": 0.00011474831718084354, + "loss": 0.2281, + "step": 33120 + }, + { + "epoch": 1.2791999691107765, + "grad_norm": 2.1078782081604004, + "learning_rate": 0.00011472257616124175, + "loss": 0.2836, + "step": 33130 + }, + { + "epoch": 1.2795860844048033, + "grad_norm": 0.19830390810966492, + "learning_rate": 0.00011469683514163995, + "loss": 0.388, + "step": 33140 + }, + { + "epoch": 1.27997219969883, + "grad_norm": 0.17538850009441376, + "learning_rate": 0.00011467109412203818, + "loss": 0.3274, + "step": 33150 + }, + { + "epoch": 1.2803583149928568, + "grad_norm": 0.7402139902114868, + "learning_rate": 0.00011464535310243641, + "loss": 0.1979, + "step": 33160 + }, + { + "epoch": 1.2807444302868838, + "grad_norm": 0.2097146362066269, + "learning_rate": 0.00011461961208283461, + "loss": 0.2464, + "step": 33170 + }, + { + "epoch": 1.2811305455809103, + "grad_norm": 1.2441083192825317, + "learning_rate": 0.00011459387106323283, + "loss": 0.266, + "step": 33180 + }, + { + "epoch": 1.2815166608749373, + "grad_norm": 2.518852710723877, + "learning_rate": 0.00011456813004363103, + "loss": 0.253, + "step": 33190 + }, + { + "epoch": 1.281902776168964, + "grad_norm": 0.8078998327255249, + "learning_rate": 0.00011454238902402925, + "loss": 0.2361, + "step": 33200 + }, + { + "epoch": 1.2822888914629909, + "grad_norm": 1.2297371625900269, + "learning_rate": 0.00011451664800442745, + "loss": 0.1974, + "step": 33210 + }, + { + "epoch": 1.2826750067570176, + "grad_norm": 0.4303855895996094, + "learning_rate": 0.00011449090698482567, + "loss": 0.3563, + "step": 33220 + }, + { + "epoch": 1.2830611220510444, + "grad_norm": 1.3215210437774658, + "learning_rate": 0.0001144651659652239, + "loss": 0.2818, + "step": 33230 + }, + { + "epoch": 1.2834472373450712, + "grad_norm": 1.546265959739685, + "learning_rate": 0.0001144394249456221, + "loss": 0.5778, + "step": 33240 + }, + { + "epoch": 1.283833352639098, + "grad_norm": 0.8895953297615051, + "learning_rate": 0.00011441368392602033, + "loss": 0.2485, + "step": 33250 + }, + { + "epoch": 1.284219467933125, + "grad_norm": 0.7534870505332947, + "learning_rate": 0.00011438794290641853, + "loss": 0.2825, + "step": 33260 + }, + { + "epoch": 1.2846055832271517, + "grad_norm": 0.052820973098278046, + "learning_rate": 0.00011436220188681674, + "loss": 0.2191, + "step": 33270 + }, + { + "epoch": 1.2849916985211784, + "grad_norm": 0.9264475107192993, + "learning_rate": 0.00011433646086721494, + "loss": 0.181, + "step": 33280 + }, + { + "epoch": 1.2853778138152052, + "grad_norm": 0.2128441333770752, + "learning_rate": 0.00011431071984761317, + "loss": 0.1819, + "step": 33290 + }, + { + "epoch": 1.285763929109232, + "grad_norm": 0.5400950312614441, + "learning_rate": 0.0001142849788280114, + "loss": 0.4316, + "step": 33300 + }, + { + "epoch": 1.2861500444032588, + "grad_norm": 0.8033271431922913, + "learning_rate": 0.00011425923780840959, + "loss": 0.2146, + "step": 33310 + }, + { + "epoch": 1.2865361596972855, + "grad_norm": 2.012575149536133, + "learning_rate": 0.00011423349678880782, + "loss": 0.4335, + "step": 33320 + }, + { + "epoch": 1.2869222749913125, + "grad_norm": 0.7352376580238342, + "learning_rate": 0.00011420775576920602, + "loss": 0.2124, + "step": 33330 + }, + { + "epoch": 1.2873083902853393, + "grad_norm": 0.769036591053009, + "learning_rate": 0.00011418201474960423, + "loss": 0.3602, + "step": 33340 + }, + { + "epoch": 1.287694505579366, + "grad_norm": 0.250592976808548, + "learning_rate": 0.00011415627373000246, + "loss": 0.1692, + "step": 33350 + }, + { + "epoch": 1.2880806208733928, + "grad_norm": 2.43820858001709, + "learning_rate": 0.00011413053271040066, + "loss": 0.2777, + "step": 33360 + }, + { + "epoch": 1.2884667361674196, + "grad_norm": 1.3179954290390015, + "learning_rate": 0.00011410479169079889, + "loss": 0.1794, + "step": 33370 + }, + { + "epoch": 1.2888528514614463, + "grad_norm": 1.0040466785430908, + "learning_rate": 0.00011407905067119709, + "loss": 0.3037, + "step": 33380 + }, + { + "epoch": 1.2892389667554731, + "grad_norm": 5.296288013458252, + "learning_rate": 0.00011405330965159531, + "loss": 0.2904, + "step": 33390 + }, + { + "epoch": 1.2896250820495, + "grad_norm": 0.4267273247241974, + "learning_rate": 0.00011402756863199351, + "loss": 0.2263, + "step": 33400 + }, + { + "epoch": 1.2900111973435266, + "grad_norm": 0.8817713260650635, + "learning_rate": 0.00011400182761239173, + "loss": 0.2715, + "step": 33410 + }, + { + "epoch": 1.2903973126375536, + "grad_norm": 2.7891275882720947, + "learning_rate": 0.00011397608659278995, + "loss": 0.2781, + "step": 33420 + }, + { + "epoch": 1.2907834279315804, + "grad_norm": 0.3013952374458313, + "learning_rate": 0.00011395034557318815, + "loss": 0.2563, + "step": 33430 + }, + { + "epoch": 1.2911695432256072, + "grad_norm": 1.766413927078247, + "learning_rate": 0.00011392460455358638, + "loss": 0.1854, + "step": 33440 + }, + { + "epoch": 1.291555658519634, + "grad_norm": 0.25331103801727295, + "learning_rate": 0.00011389886353398458, + "loss": 0.1206, + "step": 33450 + }, + { + "epoch": 1.2919417738136607, + "grad_norm": 0.036400288343429565, + "learning_rate": 0.0001138731225143828, + "loss": 0.4707, + "step": 33460 + }, + { + "epoch": 1.2923278891076875, + "grad_norm": 1.5620888471603394, + "learning_rate": 0.00011384738149478102, + "loss": 0.3312, + "step": 33470 + }, + { + "epoch": 1.2927140044017142, + "grad_norm": 0.6670392155647278, + "learning_rate": 0.00011382164047517922, + "loss": 0.2341, + "step": 33480 + }, + { + "epoch": 1.2931001196957412, + "grad_norm": 2.3108737468719482, + "learning_rate": 0.00011379589945557745, + "loss": 0.3843, + "step": 33490 + }, + { + "epoch": 1.293486234989768, + "grad_norm": 0.8025147318840027, + "learning_rate": 0.00011377015843597565, + "loss": 0.1982, + "step": 33500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.040008933376e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d96adb03b0618eda8cee0a8eda48ea362bc5f6b7 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e9c7f11c10f740ca3894e93cf549bc13beb4b2b8fd31e35fc5ce0ccf7efc040 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..3c0663c91c1bd5e5a186867cc2e3239aa6e7bedc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:909e2f55fb730af6e3b42de60bd79a9574ea1eb340e50ea9dd1290eb30e69daf +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..d15c16388ce1e7733fe120614ab1c46106d5ab08 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..ddc63aeb9309ac1f9441ee53d8fbbddb86228f12 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..1efd9cd05cedb5afc57cf3e594c32d50b8dbdb76 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/trainer_state.json @@ -0,0 +1,23834 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.3127919996911077, + "eval_steps": 500, + "global_step": 34000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + }, + { + "epoch": 1.274566585582455, + "grad_norm": 0.5019121766090393, + "learning_rate": 0.0001150314683964632, + "loss": 0.1857, + "step": 33010 + }, + { + "epoch": 1.2749527008764816, + "grad_norm": 0.258952260017395, + "learning_rate": 0.0001150057273768614, + "loss": 0.11, + "step": 33020 + }, + { + "epoch": 1.2753388161705086, + "grad_norm": 0.8540348410606384, + "learning_rate": 0.00011497998635725962, + "loss": 0.1852, + "step": 33030 + }, + { + "epoch": 1.2757249314645354, + "grad_norm": 0.08689398318529129, + "learning_rate": 0.00011495424533765783, + "loss": 0.2225, + "step": 33040 + }, + { + "epoch": 1.2761110467585621, + "grad_norm": 0.42253822088241577, + "learning_rate": 0.00011492850431805605, + "loss": 0.0751, + "step": 33050 + }, + { + "epoch": 1.276497162052589, + "grad_norm": 1.2964017391204834, + "learning_rate": 0.00011490276329845426, + "loss": 0.2384, + "step": 33060 + }, + { + "epoch": 1.2768832773466157, + "grad_norm": 0.5337836146354675, + "learning_rate": 0.00011487702227885246, + "loss": 0.1415, + "step": 33070 + }, + { + "epoch": 1.2772693926406424, + "grad_norm": 2.7771682739257812, + "learning_rate": 0.00011485128125925069, + "loss": 0.378, + "step": 33080 + }, + { + "epoch": 1.2776555079346692, + "grad_norm": 1.5107232332229614, + "learning_rate": 0.00011482554023964889, + "loss": 0.2482, + "step": 33090 + }, + { + "epoch": 1.2780416232286962, + "grad_norm": 0.6885499358177185, + "learning_rate": 0.00011479979922004711, + "loss": 0.2122, + "step": 33100 + }, + { + "epoch": 1.278427738522723, + "grad_norm": 0.9016557335853577, + "learning_rate": 0.00011477405820044533, + "loss": 0.2841, + "step": 33110 + }, + { + "epoch": 1.2788138538167497, + "grad_norm": 1.9532525539398193, + "learning_rate": 0.00011474831718084354, + "loss": 0.2281, + "step": 33120 + }, + { + "epoch": 1.2791999691107765, + "grad_norm": 2.1078782081604004, + "learning_rate": 0.00011472257616124175, + "loss": 0.2836, + "step": 33130 + }, + { + "epoch": 1.2795860844048033, + "grad_norm": 0.19830390810966492, + "learning_rate": 0.00011469683514163995, + "loss": 0.388, + "step": 33140 + }, + { + "epoch": 1.27997219969883, + "grad_norm": 0.17538850009441376, + "learning_rate": 0.00011467109412203818, + "loss": 0.3274, + "step": 33150 + }, + { + "epoch": 1.2803583149928568, + "grad_norm": 0.7402139902114868, + "learning_rate": 0.00011464535310243641, + "loss": 0.1979, + "step": 33160 + }, + { + "epoch": 1.2807444302868838, + "grad_norm": 0.2097146362066269, + "learning_rate": 0.00011461961208283461, + "loss": 0.2464, + "step": 33170 + }, + { + "epoch": 1.2811305455809103, + "grad_norm": 1.2441083192825317, + "learning_rate": 0.00011459387106323283, + "loss": 0.266, + "step": 33180 + }, + { + "epoch": 1.2815166608749373, + "grad_norm": 2.518852710723877, + "learning_rate": 0.00011456813004363103, + "loss": 0.253, + "step": 33190 + }, + { + "epoch": 1.281902776168964, + "grad_norm": 0.8078998327255249, + "learning_rate": 0.00011454238902402925, + "loss": 0.2361, + "step": 33200 + }, + { + "epoch": 1.2822888914629909, + "grad_norm": 1.2297371625900269, + "learning_rate": 0.00011451664800442745, + "loss": 0.1974, + "step": 33210 + }, + { + "epoch": 1.2826750067570176, + "grad_norm": 0.4303855895996094, + "learning_rate": 0.00011449090698482567, + "loss": 0.3563, + "step": 33220 + }, + { + "epoch": 1.2830611220510444, + "grad_norm": 1.3215210437774658, + "learning_rate": 0.0001144651659652239, + "loss": 0.2818, + "step": 33230 + }, + { + "epoch": 1.2834472373450712, + "grad_norm": 1.546265959739685, + "learning_rate": 0.0001144394249456221, + "loss": 0.5778, + "step": 33240 + }, + { + "epoch": 1.283833352639098, + "grad_norm": 0.8895953297615051, + "learning_rate": 0.00011441368392602033, + "loss": 0.2485, + "step": 33250 + }, + { + "epoch": 1.284219467933125, + "grad_norm": 0.7534870505332947, + "learning_rate": 0.00011438794290641853, + "loss": 0.2825, + "step": 33260 + }, + { + "epoch": 1.2846055832271517, + "grad_norm": 0.052820973098278046, + "learning_rate": 0.00011436220188681674, + "loss": 0.2191, + "step": 33270 + }, + { + "epoch": 1.2849916985211784, + "grad_norm": 0.9264475107192993, + "learning_rate": 0.00011433646086721494, + "loss": 0.181, + "step": 33280 + }, + { + "epoch": 1.2853778138152052, + "grad_norm": 0.2128441333770752, + "learning_rate": 0.00011431071984761317, + "loss": 0.1819, + "step": 33290 + }, + { + "epoch": 1.285763929109232, + "grad_norm": 0.5400950312614441, + "learning_rate": 0.0001142849788280114, + "loss": 0.4316, + "step": 33300 + }, + { + "epoch": 1.2861500444032588, + "grad_norm": 0.8033271431922913, + "learning_rate": 0.00011425923780840959, + "loss": 0.2146, + "step": 33310 + }, + { + "epoch": 1.2865361596972855, + "grad_norm": 2.012575149536133, + "learning_rate": 0.00011423349678880782, + "loss": 0.4335, + "step": 33320 + }, + { + "epoch": 1.2869222749913125, + "grad_norm": 0.7352376580238342, + "learning_rate": 0.00011420775576920602, + "loss": 0.2124, + "step": 33330 + }, + { + "epoch": 1.2873083902853393, + "grad_norm": 0.769036591053009, + "learning_rate": 0.00011418201474960423, + "loss": 0.3602, + "step": 33340 + }, + { + "epoch": 1.287694505579366, + "grad_norm": 0.250592976808548, + "learning_rate": 0.00011415627373000246, + "loss": 0.1692, + "step": 33350 + }, + { + "epoch": 1.2880806208733928, + "grad_norm": 2.43820858001709, + "learning_rate": 0.00011413053271040066, + "loss": 0.2777, + "step": 33360 + }, + { + "epoch": 1.2884667361674196, + "grad_norm": 1.3179954290390015, + "learning_rate": 0.00011410479169079889, + "loss": 0.1794, + "step": 33370 + }, + { + "epoch": 1.2888528514614463, + "grad_norm": 1.0040466785430908, + "learning_rate": 0.00011407905067119709, + "loss": 0.3037, + "step": 33380 + }, + { + "epoch": 1.2892389667554731, + "grad_norm": 5.296288013458252, + "learning_rate": 0.00011405330965159531, + "loss": 0.2904, + "step": 33390 + }, + { + "epoch": 1.2896250820495, + "grad_norm": 0.4267273247241974, + "learning_rate": 0.00011402756863199351, + "loss": 0.2263, + "step": 33400 + }, + { + "epoch": 1.2900111973435266, + "grad_norm": 0.8817713260650635, + "learning_rate": 0.00011400182761239173, + "loss": 0.2715, + "step": 33410 + }, + { + "epoch": 1.2903973126375536, + "grad_norm": 2.7891275882720947, + "learning_rate": 0.00011397608659278995, + "loss": 0.2781, + "step": 33420 + }, + { + "epoch": 1.2907834279315804, + "grad_norm": 0.3013952374458313, + "learning_rate": 0.00011395034557318815, + "loss": 0.2563, + "step": 33430 + }, + { + "epoch": 1.2911695432256072, + "grad_norm": 1.766413927078247, + "learning_rate": 0.00011392460455358638, + "loss": 0.1854, + "step": 33440 + }, + { + "epoch": 1.291555658519634, + "grad_norm": 0.25331103801727295, + "learning_rate": 0.00011389886353398458, + "loss": 0.1206, + "step": 33450 + }, + { + "epoch": 1.2919417738136607, + "grad_norm": 0.036400288343429565, + "learning_rate": 0.0001138731225143828, + "loss": 0.4707, + "step": 33460 + }, + { + "epoch": 1.2923278891076875, + "grad_norm": 1.5620888471603394, + "learning_rate": 0.00011384738149478102, + "loss": 0.3312, + "step": 33470 + }, + { + "epoch": 1.2927140044017142, + "grad_norm": 0.6670392155647278, + "learning_rate": 0.00011382164047517922, + "loss": 0.2341, + "step": 33480 + }, + { + "epoch": 1.2931001196957412, + "grad_norm": 2.3108737468719482, + "learning_rate": 0.00011379589945557745, + "loss": 0.3843, + "step": 33490 + }, + { + "epoch": 1.293486234989768, + "grad_norm": 0.8025147318840027, + "learning_rate": 0.00011377015843597565, + "loss": 0.1982, + "step": 33500 + }, + { + "epoch": 1.2938723502837948, + "grad_norm": 1.7835719585418701, + "learning_rate": 0.00011374441741637387, + "loss": 0.3285, + "step": 33510 + }, + { + "epoch": 1.2942584655778215, + "grad_norm": 2.041508913040161, + "learning_rate": 0.00011371867639677207, + "loss": 0.2044, + "step": 33520 + }, + { + "epoch": 1.2946445808718483, + "grad_norm": 1.103378415107727, + "learning_rate": 0.0001136929353771703, + "loss": 0.1682, + "step": 33530 + }, + { + "epoch": 1.295030696165875, + "grad_norm": 0.057376351207494736, + "learning_rate": 0.00011366719435756851, + "loss": 0.1642, + "step": 33540 + }, + { + "epoch": 1.2954168114599018, + "grad_norm": 0.6539410948753357, + "learning_rate": 0.00011364145333796671, + "loss": 0.1549, + "step": 33550 + }, + { + "epoch": 1.2958029267539288, + "grad_norm": 1.250543236732483, + "learning_rate": 0.00011361571231836494, + "loss": 0.3764, + "step": 33560 + }, + { + "epoch": 1.2961890420479556, + "grad_norm": 0.23697887361049652, + "learning_rate": 0.00011358997129876314, + "loss": 0.3999, + "step": 33570 + }, + { + "epoch": 1.2965751573419824, + "grad_norm": 0.9318505525588989, + "learning_rate": 0.00011356423027916137, + "loss": 0.4156, + "step": 33580 + }, + { + "epoch": 1.2969612726360091, + "grad_norm": 1.3910777568817139, + "learning_rate": 0.00011353848925955957, + "loss": 0.3455, + "step": 33590 + }, + { + "epoch": 1.297347387930036, + "grad_norm": 1.6764451265335083, + "learning_rate": 0.00011351274823995779, + "loss": 0.1884, + "step": 33600 + }, + { + "epoch": 1.2977335032240627, + "grad_norm": 0.9300051927566528, + "learning_rate": 0.000113487007220356, + "loss": 0.119, + "step": 33610 + }, + { + "epoch": 1.2981196185180894, + "grad_norm": 2.447462558746338, + "learning_rate": 0.00011346126620075422, + "loss": 0.4403, + "step": 33620 + }, + { + "epoch": 1.2985057338121164, + "grad_norm": 1.216407060623169, + "learning_rate": 0.00011343552518115243, + "loss": 0.2415, + "step": 33630 + }, + { + "epoch": 1.298891849106143, + "grad_norm": 2.968648910522461, + "learning_rate": 0.00011340978416155063, + "loss": 0.2899, + "step": 33640 + }, + { + "epoch": 1.29927796440017, + "grad_norm": 0.6649970412254333, + "learning_rate": 0.00011338404314194886, + "loss": 0.3809, + "step": 33650 + }, + { + "epoch": 1.2996640796941967, + "grad_norm": 1.7277917861938477, + "learning_rate": 0.00011335830212234709, + "loss": 0.3308, + "step": 33660 + }, + { + "epoch": 1.3000501949882235, + "grad_norm": 1.3269709348678589, + "learning_rate": 0.00011333256110274529, + "loss": 0.3682, + "step": 33670 + }, + { + "epoch": 1.3004363102822503, + "grad_norm": 0.20609407126903534, + "learning_rate": 0.0001133068200831435, + "loss": 0.1379, + "step": 33680 + }, + { + "epoch": 1.300822425576277, + "grad_norm": 0.6592215299606323, + "learning_rate": 0.00011328107906354171, + "loss": 0.2746, + "step": 33690 + }, + { + "epoch": 1.3012085408703038, + "grad_norm": 1.903635859489441, + "learning_rate": 0.00011325533804393993, + "loss": 0.4729, + "step": 33700 + }, + { + "epoch": 1.3015946561643306, + "grad_norm": 0.8432504534721375, + "learning_rate": 0.00011322959702433813, + "loss": 0.2835, + "step": 33710 + }, + { + "epoch": 1.3019807714583576, + "grad_norm": 0.9862542152404785, + "learning_rate": 0.00011320385600473635, + "loss": 0.1823, + "step": 33720 + }, + { + "epoch": 1.3023668867523843, + "grad_norm": 3.845738649368286, + "learning_rate": 0.00011317811498513458, + "loss": 0.2542, + "step": 33730 + }, + { + "epoch": 1.302753002046411, + "grad_norm": 0.6317747235298157, + "learning_rate": 0.00011315237396553278, + "loss": 0.22, + "step": 33740 + }, + { + "epoch": 1.3031391173404379, + "grad_norm": 2.5221354961395264, + "learning_rate": 0.000113126632945931, + "loss": 0.2253, + "step": 33750 + }, + { + "epoch": 1.3035252326344646, + "grad_norm": 1.3326247930526733, + "learning_rate": 0.0001131008919263292, + "loss": 0.2021, + "step": 33760 + }, + { + "epoch": 1.3039113479284914, + "grad_norm": 1.148047685623169, + "learning_rate": 0.00011307515090672742, + "loss": 0.3987, + "step": 33770 + }, + { + "epoch": 1.3042974632225182, + "grad_norm": 0.19721268117427826, + "learning_rate": 0.00011304940988712562, + "loss": 0.2642, + "step": 33780 + }, + { + "epoch": 1.3046835785165452, + "grad_norm": 1.4060617685317993, + "learning_rate": 0.00011302366886752385, + "loss": 0.2736, + "step": 33790 + }, + { + "epoch": 1.305069693810572, + "grad_norm": 1.0736548900604248, + "learning_rate": 0.00011299792784792207, + "loss": 0.2448, + "step": 33800 + }, + { + "epoch": 1.3054558091045987, + "grad_norm": 4.352476119995117, + "learning_rate": 0.00011297218682832027, + "loss": 0.383, + "step": 33810 + }, + { + "epoch": 1.3058419243986255, + "grad_norm": 0.2249228060245514, + "learning_rate": 0.0001129464458087185, + "loss": 0.14, + "step": 33820 + }, + { + "epoch": 1.3062280396926522, + "grad_norm": 0.4820781946182251, + "learning_rate": 0.0001129207047891167, + "loss": 0.248, + "step": 33830 + }, + { + "epoch": 1.306614154986679, + "grad_norm": 2.2983391284942627, + "learning_rate": 0.00011289496376951491, + "loss": 0.2608, + "step": 33840 + }, + { + "epoch": 1.3070002702807058, + "grad_norm": 1.3315671682357788, + "learning_rate": 0.00011286922274991314, + "loss": 0.1727, + "step": 33850 + }, + { + "epoch": 1.3073863855747327, + "grad_norm": 2.060299873352051, + "learning_rate": 0.00011284348173031134, + "loss": 0.3097, + "step": 33860 + }, + { + "epoch": 1.3077725008687593, + "grad_norm": 2.096285581588745, + "learning_rate": 0.00011281774071070957, + "loss": 0.2305, + "step": 33870 + }, + { + "epoch": 1.3081586161627863, + "grad_norm": 0.4997636675834656, + "learning_rate": 0.00011279199969110777, + "loss": 0.1993, + "step": 33880 + }, + { + "epoch": 1.308544731456813, + "grad_norm": 0.58636474609375, + "learning_rate": 0.00011276625867150599, + "loss": 0.1873, + "step": 33890 + }, + { + "epoch": 1.3089308467508398, + "grad_norm": 0.9128592610359192, + "learning_rate": 0.00011274051765190419, + "loss": 0.1885, + "step": 33900 + }, + { + "epoch": 1.3093169620448666, + "grad_norm": 2.228043794631958, + "learning_rate": 0.0001127147766323024, + "loss": 0.3649, + "step": 33910 + }, + { + "epoch": 1.3097030773388934, + "grad_norm": 1.069002389907837, + "learning_rate": 0.00011268903561270063, + "loss": 0.5454, + "step": 33920 + }, + { + "epoch": 1.3100891926329201, + "grad_norm": 0.6207597851753235, + "learning_rate": 0.00011266329459309883, + "loss": 0.2329, + "step": 33930 + }, + { + "epoch": 1.3104753079269469, + "grad_norm": 1.262247920036316, + "learning_rate": 0.00011263755357349706, + "loss": 0.3437, + "step": 33940 + }, + { + "epoch": 1.3108614232209739, + "grad_norm": 1.7429994344711304, + "learning_rate": 0.00011261181255389526, + "loss": 0.228, + "step": 33950 + }, + { + "epoch": 1.3112475385150006, + "grad_norm": 0.646900475025177, + "learning_rate": 0.00011258607153429349, + "loss": 0.3739, + "step": 33960 + }, + { + "epoch": 1.3116336538090274, + "grad_norm": 1.8228782415390015, + "learning_rate": 0.0001125603305146917, + "loss": 0.2325, + "step": 33970 + }, + { + "epoch": 1.3120197691030542, + "grad_norm": 3.539228916168213, + "learning_rate": 0.0001125345894950899, + "loss": 0.194, + "step": 33980 + }, + { + "epoch": 1.312405884397081, + "grad_norm": 1.2801135778427124, + "learning_rate": 0.00011250884847548813, + "loss": 0.3069, + "step": 33990 + }, + { + "epoch": 1.3127919996911077, + "grad_norm": 3.6265695095062256, + "learning_rate": 0.00011248310745588633, + "loss": 0.3113, + "step": 34000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.070456827904e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c0a7c75e4cd7db66d8895a2774ae55f6763b5754 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac8b3492214f38124166691dd3874eab4f3337125bacd02b92e1dc8e8e6144f8 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..249411b3dffc8b6b216e5ff399eba31e4b77e32a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cfca8b852809ef3a0477c813aa8e0e45ee0f5e1e6414e2f80a60cda896e7ffb1 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..d67a270424b5eb9fad337b61918680848aa1ed60 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..3eb0546c54535e4045b73492730a43dbb26f46da Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..e76daa4372fceb2ae203dd1f512da42e3edfb66d --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/trainer_state.json @@ -0,0 +1,24184 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.3320977643924476, + "eval_steps": 500, + "global_step": 34500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + }, + { + "epoch": 1.274566585582455, + "grad_norm": 0.5019121766090393, + "learning_rate": 0.0001150314683964632, + "loss": 0.1857, + "step": 33010 + }, + { + "epoch": 1.2749527008764816, + "grad_norm": 0.258952260017395, + "learning_rate": 0.0001150057273768614, + "loss": 0.11, + "step": 33020 + }, + { + "epoch": 1.2753388161705086, + "grad_norm": 0.8540348410606384, + "learning_rate": 0.00011497998635725962, + "loss": 0.1852, + "step": 33030 + }, + { + "epoch": 1.2757249314645354, + "grad_norm": 0.08689398318529129, + "learning_rate": 0.00011495424533765783, + "loss": 0.2225, + "step": 33040 + }, + { + "epoch": 1.2761110467585621, + "grad_norm": 0.42253822088241577, + "learning_rate": 0.00011492850431805605, + "loss": 0.0751, + "step": 33050 + }, + { + "epoch": 1.276497162052589, + "grad_norm": 1.2964017391204834, + "learning_rate": 0.00011490276329845426, + "loss": 0.2384, + "step": 33060 + }, + { + "epoch": 1.2768832773466157, + "grad_norm": 0.5337836146354675, + "learning_rate": 0.00011487702227885246, + "loss": 0.1415, + "step": 33070 + }, + { + "epoch": 1.2772693926406424, + "grad_norm": 2.7771682739257812, + "learning_rate": 0.00011485128125925069, + "loss": 0.378, + "step": 33080 + }, + { + "epoch": 1.2776555079346692, + "grad_norm": 1.5107232332229614, + "learning_rate": 0.00011482554023964889, + "loss": 0.2482, + "step": 33090 + }, + { + "epoch": 1.2780416232286962, + "grad_norm": 0.6885499358177185, + "learning_rate": 0.00011479979922004711, + "loss": 0.2122, + "step": 33100 + }, + { + "epoch": 1.278427738522723, + "grad_norm": 0.9016557335853577, + "learning_rate": 0.00011477405820044533, + "loss": 0.2841, + "step": 33110 + }, + { + "epoch": 1.2788138538167497, + "grad_norm": 1.9532525539398193, + "learning_rate": 0.00011474831718084354, + "loss": 0.2281, + "step": 33120 + }, + { + "epoch": 1.2791999691107765, + "grad_norm": 2.1078782081604004, + "learning_rate": 0.00011472257616124175, + "loss": 0.2836, + "step": 33130 + }, + { + "epoch": 1.2795860844048033, + "grad_norm": 0.19830390810966492, + "learning_rate": 0.00011469683514163995, + "loss": 0.388, + "step": 33140 + }, + { + "epoch": 1.27997219969883, + "grad_norm": 0.17538850009441376, + "learning_rate": 0.00011467109412203818, + "loss": 0.3274, + "step": 33150 + }, + { + "epoch": 1.2803583149928568, + "grad_norm": 0.7402139902114868, + "learning_rate": 0.00011464535310243641, + "loss": 0.1979, + "step": 33160 + }, + { + "epoch": 1.2807444302868838, + "grad_norm": 0.2097146362066269, + "learning_rate": 0.00011461961208283461, + "loss": 0.2464, + "step": 33170 + }, + { + "epoch": 1.2811305455809103, + "grad_norm": 1.2441083192825317, + "learning_rate": 0.00011459387106323283, + "loss": 0.266, + "step": 33180 + }, + { + "epoch": 1.2815166608749373, + "grad_norm": 2.518852710723877, + "learning_rate": 0.00011456813004363103, + "loss": 0.253, + "step": 33190 + }, + { + "epoch": 1.281902776168964, + "grad_norm": 0.8078998327255249, + "learning_rate": 0.00011454238902402925, + "loss": 0.2361, + "step": 33200 + }, + { + "epoch": 1.2822888914629909, + "grad_norm": 1.2297371625900269, + "learning_rate": 0.00011451664800442745, + "loss": 0.1974, + "step": 33210 + }, + { + "epoch": 1.2826750067570176, + "grad_norm": 0.4303855895996094, + "learning_rate": 0.00011449090698482567, + "loss": 0.3563, + "step": 33220 + }, + { + "epoch": 1.2830611220510444, + "grad_norm": 1.3215210437774658, + "learning_rate": 0.0001144651659652239, + "loss": 0.2818, + "step": 33230 + }, + { + "epoch": 1.2834472373450712, + "grad_norm": 1.546265959739685, + "learning_rate": 0.0001144394249456221, + "loss": 0.5778, + "step": 33240 + }, + { + "epoch": 1.283833352639098, + "grad_norm": 0.8895953297615051, + "learning_rate": 0.00011441368392602033, + "loss": 0.2485, + "step": 33250 + }, + { + "epoch": 1.284219467933125, + "grad_norm": 0.7534870505332947, + "learning_rate": 0.00011438794290641853, + "loss": 0.2825, + "step": 33260 + }, + { + "epoch": 1.2846055832271517, + "grad_norm": 0.052820973098278046, + "learning_rate": 0.00011436220188681674, + "loss": 0.2191, + "step": 33270 + }, + { + "epoch": 1.2849916985211784, + "grad_norm": 0.9264475107192993, + "learning_rate": 0.00011433646086721494, + "loss": 0.181, + "step": 33280 + }, + { + "epoch": 1.2853778138152052, + "grad_norm": 0.2128441333770752, + "learning_rate": 0.00011431071984761317, + "loss": 0.1819, + "step": 33290 + }, + { + "epoch": 1.285763929109232, + "grad_norm": 0.5400950312614441, + "learning_rate": 0.0001142849788280114, + "loss": 0.4316, + "step": 33300 + }, + { + "epoch": 1.2861500444032588, + "grad_norm": 0.8033271431922913, + "learning_rate": 0.00011425923780840959, + "loss": 0.2146, + "step": 33310 + }, + { + "epoch": 1.2865361596972855, + "grad_norm": 2.012575149536133, + "learning_rate": 0.00011423349678880782, + "loss": 0.4335, + "step": 33320 + }, + { + "epoch": 1.2869222749913125, + "grad_norm": 0.7352376580238342, + "learning_rate": 0.00011420775576920602, + "loss": 0.2124, + "step": 33330 + }, + { + "epoch": 1.2873083902853393, + "grad_norm": 0.769036591053009, + "learning_rate": 0.00011418201474960423, + "loss": 0.3602, + "step": 33340 + }, + { + "epoch": 1.287694505579366, + "grad_norm": 0.250592976808548, + "learning_rate": 0.00011415627373000246, + "loss": 0.1692, + "step": 33350 + }, + { + "epoch": 1.2880806208733928, + "grad_norm": 2.43820858001709, + "learning_rate": 0.00011413053271040066, + "loss": 0.2777, + "step": 33360 + }, + { + "epoch": 1.2884667361674196, + "grad_norm": 1.3179954290390015, + "learning_rate": 0.00011410479169079889, + "loss": 0.1794, + "step": 33370 + }, + { + "epoch": 1.2888528514614463, + "grad_norm": 1.0040466785430908, + "learning_rate": 0.00011407905067119709, + "loss": 0.3037, + "step": 33380 + }, + { + "epoch": 1.2892389667554731, + "grad_norm": 5.296288013458252, + "learning_rate": 0.00011405330965159531, + "loss": 0.2904, + "step": 33390 + }, + { + "epoch": 1.2896250820495, + "grad_norm": 0.4267273247241974, + "learning_rate": 0.00011402756863199351, + "loss": 0.2263, + "step": 33400 + }, + { + "epoch": 1.2900111973435266, + "grad_norm": 0.8817713260650635, + "learning_rate": 0.00011400182761239173, + "loss": 0.2715, + "step": 33410 + }, + { + "epoch": 1.2903973126375536, + "grad_norm": 2.7891275882720947, + "learning_rate": 0.00011397608659278995, + "loss": 0.2781, + "step": 33420 + }, + { + "epoch": 1.2907834279315804, + "grad_norm": 0.3013952374458313, + "learning_rate": 0.00011395034557318815, + "loss": 0.2563, + "step": 33430 + }, + { + "epoch": 1.2911695432256072, + "grad_norm": 1.766413927078247, + "learning_rate": 0.00011392460455358638, + "loss": 0.1854, + "step": 33440 + }, + { + "epoch": 1.291555658519634, + "grad_norm": 0.25331103801727295, + "learning_rate": 0.00011389886353398458, + "loss": 0.1206, + "step": 33450 + }, + { + "epoch": 1.2919417738136607, + "grad_norm": 0.036400288343429565, + "learning_rate": 0.0001138731225143828, + "loss": 0.4707, + "step": 33460 + }, + { + "epoch": 1.2923278891076875, + "grad_norm": 1.5620888471603394, + "learning_rate": 0.00011384738149478102, + "loss": 0.3312, + "step": 33470 + }, + { + "epoch": 1.2927140044017142, + "grad_norm": 0.6670392155647278, + "learning_rate": 0.00011382164047517922, + "loss": 0.2341, + "step": 33480 + }, + { + "epoch": 1.2931001196957412, + "grad_norm": 2.3108737468719482, + "learning_rate": 0.00011379589945557745, + "loss": 0.3843, + "step": 33490 + }, + { + "epoch": 1.293486234989768, + "grad_norm": 0.8025147318840027, + "learning_rate": 0.00011377015843597565, + "loss": 0.1982, + "step": 33500 + }, + { + "epoch": 1.2938723502837948, + "grad_norm": 1.7835719585418701, + "learning_rate": 0.00011374441741637387, + "loss": 0.3285, + "step": 33510 + }, + { + "epoch": 1.2942584655778215, + "grad_norm": 2.041508913040161, + "learning_rate": 0.00011371867639677207, + "loss": 0.2044, + "step": 33520 + }, + { + "epoch": 1.2946445808718483, + "grad_norm": 1.103378415107727, + "learning_rate": 0.0001136929353771703, + "loss": 0.1682, + "step": 33530 + }, + { + "epoch": 1.295030696165875, + "grad_norm": 0.057376351207494736, + "learning_rate": 0.00011366719435756851, + "loss": 0.1642, + "step": 33540 + }, + { + "epoch": 1.2954168114599018, + "grad_norm": 0.6539410948753357, + "learning_rate": 0.00011364145333796671, + "loss": 0.1549, + "step": 33550 + }, + { + "epoch": 1.2958029267539288, + "grad_norm": 1.250543236732483, + "learning_rate": 0.00011361571231836494, + "loss": 0.3764, + "step": 33560 + }, + { + "epoch": 1.2961890420479556, + "grad_norm": 0.23697887361049652, + "learning_rate": 0.00011358997129876314, + "loss": 0.3999, + "step": 33570 + }, + { + "epoch": 1.2965751573419824, + "grad_norm": 0.9318505525588989, + "learning_rate": 0.00011356423027916137, + "loss": 0.4156, + "step": 33580 + }, + { + "epoch": 1.2969612726360091, + "grad_norm": 1.3910777568817139, + "learning_rate": 0.00011353848925955957, + "loss": 0.3455, + "step": 33590 + }, + { + "epoch": 1.297347387930036, + "grad_norm": 1.6764451265335083, + "learning_rate": 0.00011351274823995779, + "loss": 0.1884, + "step": 33600 + }, + { + "epoch": 1.2977335032240627, + "grad_norm": 0.9300051927566528, + "learning_rate": 0.000113487007220356, + "loss": 0.119, + "step": 33610 + }, + { + "epoch": 1.2981196185180894, + "grad_norm": 2.447462558746338, + "learning_rate": 0.00011346126620075422, + "loss": 0.4403, + "step": 33620 + }, + { + "epoch": 1.2985057338121164, + "grad_norm": 1.216407060623169, + "learning_rate": 0.00011343552518115243, + "loss": 0.2415, + "step": 33630 + }, + { + "epoch": 1.298891849106143, + "grad_norm": 2.968648910522461, + "learning_rate": 0.00011340978416155063, + "loss": 0.2899, + "step": 33640 + }, + { + "epoch": 1.29927796440017, + "grad_norm": 0.6649970412254333, + "learning_rate": 0.00011338404314194886, + "loss": 0.3809, + "step": 33650 + }, + { + "epoch": 1.2996640796941967, + "grad_norm": 1.7277917861938477, + "learning_rate": 0.00011335830212234709, + "loss": 0.3308, + "step": 33660 + }, + { + "epoch": 1.3000501949882235, + "grad_norm": 1.3269709348678589, + "learning_rate": 0.00011333256110274529, + "loss": 0.3682, + "step": 33670 + }, + { + "epoch": 1.3004363102822503, + "grad_norm": 0.20609407126903534, + "learning_rate": 0.0001133068200831435, + "loss": 0.1379, + "step": 33680 + }, + { + "epoch": 1.300822425576277, + "grad_norm": 0.6592215299606323, + "learning_rate": 0.00011328107906354171, + "loss": 0.2746, + "step": 33690 + }, + { + "epoch": 1.3012085408703038, + "grad_norm": 1.903635859489441, + "learning_rate": 0.00011325533804393993, + "loss": 0.4729, + "step": 33700 + }, + { + "epoch": 1.3015946561643306, + "grad_norm": 0.8432504534721375, + "learning_rate": 0.00011322959702433813, + "loss": 0.2835, + "step": 33710 + }, + { + "epoch": 1.3019807714583576, + "grad_norm": 0.9862542152404785, + "learning_rate": 0.00011320385600473635, + "loss": 0.1823, + "step": 33720 + }, + { + "epoch": 1.3023668867523843, + "grad_norm": 3.845738649368286, + "learning_rate": 0.00011317811498513458, + "loss": 0.2542, + "step": 33730 + }, + { + "epoch": 1.302753002046411, + "grad_norm": 0.6317747235298157, + "learning_rate": 0.00011315237396553278, + "loss": 0.22, + "step": 33740 + }, + { + "epoch": 1.3031391173404379, + "grad_norm": 2.5221354961395264, + "learning_rate": 0.000113126632945931, + "loss": 0.2253, + "step": 33750 + }, + { + "epoch": 1.3035252326344646, + "grad_norm": 1.3326247930526733, + "learning_rate": 0.0001131008919263292, + "loss": 0.2021, + "step": 33760 + }, + { + "epoch": 1.3039113479284914, + "grad_norm": 1.148047685623169, + "learning_rate": 0.00011307515090672742, + "loss": 0.3987, + "step": 33770 + }, + { + "epoch": 1.3042974632225182, + "grad_norm": 0.19721268117427826, + "learning_rate": 0.00011304940988712562, + "loss": 0.2642, + "step": 33780 + }, + { + "epoch": 1.3046835785165452, + "grad_norm": 1.4060617685317993, + "learning_rate": 0.00011302366886752385, + "loss": 0.2736, + "step": 33790 + }, + { + "epoch": 1.305069693810572, + "grad_norm": 1.0736548900604248, + "learning_rate": 0.00011299792784792207, + "loss": 0.2448, + "step": 33800 + }, + { + "epoch": 1.3054558091045987, + "grad_norm": 4.352476119995117, + "learning_rate": 0.00011297218682832027, + "loss": 0.383, + "step": 33810 + }, + { + "epoch": 1.3058419243986255, + "grad_norm": 0.2249228060245514, + "learning_rate": 0.0001129464458087185, + "loss": 0.14, + "step": 33820 + }, + { + "epoch": 1.3062280396926522, + "grad_norm": 0.4820781946182251, + "learning_rate": 0.0001129207047891167, + "loss": 0.248, + "step": 33830 + }, + { + "epoch": 1.306614154986679, + "grad_norm": 2.2983391284942627, + "learning_rate": 0.00011289496376951491, + "loss": 0.2608, + "step": 33840 + }, + { + "epoch": 1.3070002702807058, + "grad_norm": 1.3315671682357788, + "learning_rate": 0.00011286922274991314, + "loss": 0.1727, + "step": 33850 + }, + { + "epoch": 1.3073863855747327, + "grad_norm": 2.060299873352051, + "learning_rate": 0.00011284348173031134, + "loss": 0.3097, + "step": 33860 + }, + { + "epoch": 1.3077725008687593, + "grad_norm": 2.096285581588745, + "learning_rate": 0.00011281774071070957, + "loss": 0.2305, + "step": 33870 + }, + { + "epoch": 1.3081586161627863, + "grad_norm": 0.4997636675834656, + "learning_rate": 0.00011279199969110777, + "loss": 0.1993, + "step": 33880 + }, + { + "epoch": 1.308544731456813, + "grad_norm": 0.58636474609375, + "learning_rate": 0.00011276625867150599, + "loss": 0.1873, + "step": 33890 + }, + { + "epoch": 1.3089308467508398, + "grad_norm": 0.9128592610359192, + "learning_rate": 0.00011274051765190419, + "loss": 0.1885, + "step": 33900 + }, + { + "epoch": 1.3093169620448666, + "grad_norm": 2.228043794631958, + "learning_rate": 0.0001127147766323024, + "loss": 0.3649, + "step": 33910 + }, + { + "epoch": 1.3097030773388934, + "grad_norm": 1.069002389907837, + "learning_rate": 0.00011268903561270063, + "loss": 0.5454, + "step": 33920 + }, + { + "epoch": 1.3100891926329201, + "grad_norm": 0.6207597851753235, + "learning_rate": 0.00011266329459309883, + "loss": 0.2329, + "step": 33930 + }, + { + "epoch": 1.3104753079269469, + "grad_norm": 1.262247920036316, + "learning_rate": 0.00011263755357349706, + "loss": 0.3437, + "step": 33940 + }, + { + "epoch": 1.3108614232209739, + "grad_norm": 1.7429994344711304, + "learning_rate": 0.00011261181255389526, + "loss": 0.228, + "step": 33950 + }, + { + "epoch": 1.3112475385150006, + "grad_norm": 0.646900475025177, + "learning_rate": 0.00011258607153429349, + "loss": 0.3739, + "step": 33960 + }, + { + "epoch": 1.3116336538090274, + "grad_norm": 1.8228782415390015, + "learning_rate": 0.0001125603305146917, + "loss": 0.2325, + "step": 33970 + }, + { + "epoch": 1.3120197691030542, + "grad_norm": 3.539228916168213, + "learning_rate": 0.0001125345894950899, + "loss": 0.194, + "step": 33980 + }, + { + "epoch": 1.312405884397081, + "grad_norm": 1.2801135778427124, + "learning_rate": 0.00011250884847548813, + "loss": 0.3069, + "step": 33990 + }, + { + "epoch": 1.3127919996911077, + "grad_norm": 3.6265695095062256, + "learning_rate": 0.00011248310745588633, + "loss": 0.3113, + "step": 34000 + }, + { + "epoch": 1.3131781149851345, + "grad_norm": 0.07370063662528992, + "learning_rate": 0.00011245736643628455, + "loss": 0.1449, + "step": 34010 + }, + { + "epoch": 1.3135642302791615, + "grad_norm": 1.0295637845993042, + "learning_rate": 0.00011243162541668275, + "loss": 0.22, + "step": 34020 + }, + { + "epoch": 1.313950345573188, + "grad_norm": 0.8803662061691284, + "learning_rate": 0.00011240588439708098, + "loss": 0.1368, + "step": 34030 + }, + { + "epoch": 1.314336460867215, + "grad_norm": 1.6597707271575928, + "learning_rate": 0.00011238014337747919, + "loss": 0.3038, + "step": 34040 + }, + { + "epoch": 1.3147225761612418, + "grad_norm": 2.115492343902588, + "learning_rate": 0.00011235440235787739, + "loss": 0.1754, + "step": 34050 + }, + { + "epoch": 1.3151086914552685, + "grad_norm": 0.8143919706344604, + "learning_rate": 0.00011232866133827562, + "loss": 0.3764, + "step": 34060 + }, + { + "epoch": 1.3154948067492953, + "grad_norm": 0.14369767904281616, + "learning_rate": 0.00011230292031867382, + "loss": 0.1742, + "step": 34070 + }, + { + "epoch": 1.315880922043322, + "grad_norm": 1.0129845142364502, + "learning_rate": 0.00011227717929907205, + "loss": 0.1458, + "step": 34080 + }, + { + "epoch": 1.316267037337349, + "grad_norm": 2.7300291061401367, + "learning_rate": 0.00011225143827947025, + "loss": 0.3939, + "step": 34090 + }, + { + "epoch": 1.3166531526313756, + "grad_norm": 0.20205609500408173, + "learning_rate": 0.00011222569725986847, + "loss": 0.29, + "step": 34100 + }, + { + "epoch": 1.3170392679254026, + "grad_norm": 1.8928464651107788, + "learning_rate": 0.00011219995624026669, + "loss": 0.1742, + "step": 34110 + }, + { + "epoch": 1.3174253832194294, + "grad_norm": 0.2639687955379486, + "learning_rate": 0.00011217421522066488, + "loss": 0.1745, + "step": 34120 + }, + { + "epoch": 1.3178114985134561, + "grad_norm": 0.5906389355659485, + "learning_rate": 0.00011214847420106311, + "loss": 0.2134, + "step": 34130 + }, + { + "epoch": 1.318197613807483, + "grad_norm": 0.9190629720687866, + "learning_rate": 0.00011212273318146131, + "loss": 0.2547, + "step": 34140 + }, + { + "epoch": 1.3185837291015097, + "grad_norm": 0.5749151110649109, + "learning_rate": 0.00011209699216185954, + "loss": 0.1688, + "step": 34150 + }, + { + "epoch": 1.3189698443955364, + "grad_norm": 0.82295823097229, + "learning_rate": 0.00011207125114225777, + "loss": 0.2884, + "step": 34160 + }, + { + "epoch": 1.3193559596895632, + "grad_norm": 0.07816460728645325, + "learning_rate": 0.00011204551012265597, + "loss": 0.2418, + "step": 34170 + }, + { + "epoch": 1.3197420749835902, + "grad_norm": 0.6417407393455505, + "learning_rate": 0.00011201976910305418, + "loss": 0.2557, + "step": 34180 + }, + { + "epoch": 1.320128190277617, + "grad_norm": 6.093267440795898, + "learning_rate": 0.00011199402808345238, + "loss": 0.3088, + "step": 34190 + }, + { + "epoch": 1.3205143055716437, + "grad_norm": 1.8861887454986572, + "learning_rate": 0.0001119682870638506, + "loss": 0.2204, + "step": 34200 + }, + { + "epoch": 1.3209004208656705, + "grad_norm": 2.3272714614868164, + "learning_rate": 0.0001119425460442488, + "loss": 0.2236, + "step": 34210 + }, + { + "epoch": 1.3212865361596973, + "grad_norm": 0.9608810544013977, + "learning_rate": 0.00011191680502464703, + "loss": 0.1897, + "step": 34220 + }, + { + "epoch": 1.321672651453724, + "grad_norm": 1.2157350778579712, + "learning_rate": 0.00011189106400504526, + "loss": 0.1526, + "step": 34230 + }, + { + "epoch": 1.3220587667477508, + "grad_norm": 1.6684671640396118, + "learning_rate": 0.00011186532298544346, + "loss": 0.3394, + "step": 34240 + }, + { + "epoch": 1.3224448820417778, + "grad_norm": 2.0432374477386475, + "learning_rate": 0.00011183958196584167, + "loss": 0.2183, + "step": 34250 + }, + { + "epoch": 1.3228309973358043, + "grad_norm": 0.9436892867088318, + "learning_rate": 0.00011181384094623988, + "loss": 0.2947, + "step": 34260 + }, + { + "epoch": 1.3232171126298313, + "grad_norm": 0.23260092735290527, + "learning_rate": 0.0001117880999266381, + "loss": 0.114, + "step": 34270 + }, + { + "epoch": 1.323603227923858, + "grad_norm": 1.2291594743728638, + "learning_rate": 0.0001117623589070363, + "loss": 0.3145, + "step": 34280 + }, + { + "epoch": 1.3239893432178849, + "grad_norm": 0.41411107778549194, + "learning_rate": 0.00011173661788743452, + "loss": 0.2937, + "step": 34290 + }, + { + "epoch": 1.3243754585119116, + "grad_norm": 2.354405164718628, + "learning_rate": 0.00011171087686783275, + "loss": 0.3933, + "step": 34300 + }, + { + "epoch": 1.3247615738059384, + "grad_norm": 2.6997978687286377, + "learning_rate": 0.00011168513584823095, + "loss": 0.1494, + "step": 34310 + }, + { + "epoch": 1.3251476890999654, + "grad_norm": 2.8430919647216797, + "learning_rate": 0.00011165939482862916, + "loss": 0.2869, + "step": 34320 + }, + { + "epoch": 1.325533804393992, + "grad_norm": 1.1737356185913086, + "learning_rate": 0.00011163365380902738, + "loss": 0.2792, + "step": 34330 + }, + { + "epoch": 1.325919919688019, + "grad_norm": 4.123973846435547, + "learning_rate": 0.00011160791278942559, + "loss": 0.5211, + "step": 34340 + }, + { + "epoch": 1.3263060349820457, + "grad_norm": 0.8862038850784302, + "learning_rate": 0.00011158217176982382, + "loss": 0.2976, + "step": 34350 + }, + { + "epoch": 1.3266921502760725, + "grad_norm": 1.8690590858459473, + "learning_rate": 0.00011155643075022202, + "loss": 0.2485, + "step": 34360 + }, + { + "epoch": 1.3270782655700992, + "grad_norm": 2.885589599609375, + "learning_rate": 0.00011153068973062024, + "loss": 0.3084, + "step": 34370 + }, + { + "epoch": 1.327464380864126, + "grad_norm": 0.9898788928985596, + "learning_rate": 0.00011150494871101844, + "loss": 0.261, + "step": 34380 + }, + { + "epoch": 1.3278504961581528, + "grad_norm": 0.6879653930664062, + "learning_rate": 0.00011147920769141667, + "loss": 0.2082, + "step": 34390 + }, + { + "epoch": 1.3282366114521795, + "grad_norm": 1.2619003057479858, + "learning_rate": 0.00011145346667181487, + "loss": 0.2402, + "step": 34400 + }, + { + "epoch": 1.3286227267462065, + "grad_norm": 1.1212007999420166, + "learning_rate": 0.00011142772565221308, + "loss": 0.3525, + "step": 34410 + }, + { + "epoch": 1.3290088420402333, + "grad_norm": 1.8431956768035889, + "learning_rate": 0.00011140198463261131, + "loss": 0.2212, + "step": 34420 + }, + { + "epoch": 1.32939495733426, + "grad_norm": 0.6185423731803894, + "learning_rate": 0.00011137624361300951, + "loss": 0.2455, + "step": 34430 + }, + { + "epoch": 1.3297810726282868, + "grad_norm": 2.3791301250457764, + "learning_rate": 0.00011135050259340774, + "loss": 0.1763, + "step": 34440 + }, + { + "epoch": 1.3301671879223136, + "grad_norm": 0.4928603768348694, + "learning_rate": 0.00011132476157380594, + "loss": 0.2381, + "step": 34450 + }, + { + "epoch": 1.3305533032163404, + "grad_norm": 1.5636029243469238, + "learning_rate": 0.00011129902055420416, + "loss": 0.1368, + "step": 34460 + }, + { + "epoch": 1.3309394185103671, + "grad_norm": 0.9425283074378967, + "learning_rate": 0.00011127327953460238, + "loss": 0.2825, + "step": 34470 + }, + { + "epoch": 1.3313255338043941, + "grad_norm": 1.2257115840911865, + "learning_rate": 0.00011124753851500058, + "loss": 0.2547, + "step": 34480 + }, + { + "epoch": 1.3317116490984207, + "grad_norm": 0.9416170716285706, + "learning_rate": 0.0001112217974953988, + "loss": 0.2766, + "step": 34490 + }, + { + "epoch": 1.3320977643924476, + "grad_norm": 0.5123847126960754, + "learning_rate": 0.000111196056475797, + "loss": 0.4733, + "step": 34500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.100904722432e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1177b9aee606e82030928a1f550e1cd8d20fedd3 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:043a34cd5635bb13edb6134da2cf19b1089edec15e4c1e78709fd0d4d00d0001 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..fcb6afe6bd71cb7abe8214f98e15f76ae2fe7423 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14fe2eb7ff7eb41b0112ff29d96de8e0d69570b899061b92c700ab870bfeffcf +size 291766272 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..8993df6000f3ef4f7ad21819dd6dd0bdaf113e0e Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..a6a7177715b76bcffcec5ce7f60b17d0d66ec67a Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..83e5158a5e5055715f83a85341d469f512575424 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/trainer_state.json @@ -0,0 +1,24534 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.3514035290937874, + "eval_steps": 500, + "global_step": 35000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + }, + { + "epoch": 1.274566585582455, + "grad_norm": 0.5019121766090393, + "learning_rate": 0.0001150314683964632, + "loss": 0.1857, + "step": 33010 + }, + { + "epoch": 1.2749527008764816, + "grad_norm": 0.258952260017395, + "learning_rate": 0.0001150057273768614, + "loss": 0.11, + "step": 33020 + }, + { + "epoch": 1.2753388161705086, + "grad_norm": 0.8540348410606384, + "learning_rate": 0.00011497998635725962, + "loss": 0.1852, + "step": 33030 + }, + { + "epoch": 1.2757249314645354, + "grad_norm": 0.08689398318529129, + "learning_rate": 0.00011495424533765783, + "loss": 0.2225, + "step": 33040 + }, + { + "epoch": 1.2761110467585621, + "grad_norm": 0.42253822088241577, + "learning_rate": 0.00011492850431805605, + "loss": 0.0751, + "step": 33050 + }, + { + "epoch": 1.276497162052589, + "grad_norm": 1.2964017391204834, + "learning_rate": 0.00011490276329845426, + "loss": 0.2384, + "step": 33060 + }, + { + "epoch": 1.2768832773466157, + "grad_norm": 0.5337836146354675, + "learning_rate": 0.00011487702227885246, + "loss": 0.1415, + "step": 33070 + }, + { + "epoch": 1.2772693926406424, + "grad_norm": 2.7771682739257812, + "learning_rate": 0.00011485128125925069, + "loss": 0.378, + "step": 33080 + }, + { + "epoch": 1.2776555079346692, + "grad_norm": 1.5107232332229614, + "learning_rate": 0.00011482554023964889, + "loss": 0.2482, + "step": 33090 + }, + { + "epoch": 1.2780416232286962, + "grad_norm": 0.6885499358177185, + "learning_rate": 0.00011479979922004711, + "loss": 0.2122, + "step": 33100 + }, + { + "epoch": 1.278427738522723, + "grad_norm": 0.9016557335853577, + "learning_rate": 0.00011477405820044533, + "loss": 0.2841, + "step": 33110 + }, + { + "epoch": 1.2788138538167497, + "grad_norm": 1.9532525539398193, + "learning_rate": 0.00011474831718084354, + "loss": 0.2281, + "step": 33120 + }, + { + "epoch": 1.2791999691107765, + "grad_norm": 2.1078782081604004, + "learning_rate": 0.00011472257616124175, + "loss": 0.2836, + "step": 33130 + }, + { + "epoch": 1.2795860844048033, + "grad_norm": 0.19830390810966492, + "learning_rate": 0.00011469683514163995, + "loss": 0.388, + "step": 33140 + }, + { + "epoch": 1.27997219969883, + "grad_norm": 0.17538850009441376, + "learning_rate": 0.00011467109412203818, + "loss": 0.3274, + "step": 33150 + }, + { + "epoch": 1.2803583149928568, + "grad_norm": 0.7402139902114868, + "learning_rate": 0.00011464535310243641, + "loss": 0.1979, + "step": 33160 + }, + { + "epoch": 1.2807444302868838, + "grad_norm": 0.2097146362066269, + "learning_rate": 0.00011461961208283461, + "loss": 0.2464, + "step": 33170 + }, + { + "epoch": 1.2811305455809103, + "grad_norm": 1.2441083192825317, + "learning_rate": 0.00011459387106323283, + "loss": 0.266, + "step": 33180 + }, + { + "epoch": 1.2815166608749373, + "grad_norm": 2.518852710723877, + "learning_rate": 0.00011456813004363103, + "loss": 0.253, + "step": 33190 + }, + { + "epoch": 1.281902776168964, + "grad_norm": 0.8078998327255249, + "learning_rate": 0.00011454238902402925, + "loss": 0.2361, + "step": 33200 + }, + { + "epoch": 1.2822888914629909, + "grad_norm": 1.2297371625900269, + "learning_rate": 0.00011451664800442745, + "loss": 0.1974, + "step": 33210 + }, + { + "epoch": 1.2826750067570176, + "grad_norm": 0.4303855895996094, + "learning_rate": 0.00011449090698482567, + "loss": 0.3563, + "step": 33220 + }, + { + "epoch": 1.2830611220510444, + "grad_norm": 1.3215210437774658, + "learning_rate": 0.0001144651659652239, + "loss": 0.2818, + "step": 33230 + }, + { + "epoch": 1.2834472373450712, + "grad_norm": 1.546265959739685, + "learning_rate": 0.0001144394249456221, + "loss": 0.5778, + "step": 33240 + }, + { + "epoch": 1.283833352639098, + "grad_norm": 0.8895953297615051, + "learning_rate": 0.00011441368392602033, + "loss": 0.2485, + "step": 33250 + }, + { + "epoch": 1.284219467933125, + "grad_norm": 0.7534870505332947, + "learning_rate": 0.00011438794290641853, + "loss": 0.2825, + "step": 33260 + }, + { + "epoch": 1.2846055832271517, + "grad_norm": 0.052820973098278046, + "learning_rate": 0.00011436220188681674, + "loss": 0.2191, + "step": 33270 + }, + { + "epoch": 1.2849916985211784, + "grad_norm": 0.9264475107192993, + "learning_rate": 0.00011433646086721494, + "loss": 0.181, + "step": 33280 + }, + { + "epoch": 1.2853778138152052, + "grad_norm": 0.2128441333770752, + "learning_rate": 0.00011431071984761317, + "loss": 0.1819, + "step": 33290 + }, + { + "epoch": 1.285763929109232, + "grad_norm": 0.5400950312614441, + "learning_rate": 0.0001142849788280114, + "loss": 0.4316, + "step": 33300 + }, + { + "epoch": 1.2861500444032588, + "grad_norm": 0.8033271431922913, + "learning_rate": 0.00011425923780840959, + "loss": 0.2146, + "step": 33310 + }, + { + "epoch": 1.2865361596972855, + "grad_norm": 2.012575149536133, + "learning_rate": 0.00011423349678880782, + "loss": 0.4335, + "step": 33320 + }, + { + "epoch": 1.2869222749913125, + "grad_norm": 0.7352376580238342, + "learning_rate": 0.00011420775576920602, + "loss": 0.2124, + "step": 33330 + }, + { + "epoch": 1.2873083902853393, + "grad_norm": 0.769036591053009, + "learning_rate": 0.00011418201474960423, + "loss": 0.3602, + "step": 33340 + }, + { + "epoch": 1.287694505579366, + "grad_norm": 0.250592976808548, + "learning_rate": 0.00011415627373000246, + "loss": 0.1692, + "step": 33350 + }, + { + "epoch": 1.2880806208733928, + "grad_norm": 2.43820858001709, + "learning_rate": 0.00011413053271040066, + "loss": 0.2777, + "step": 33360 + }, + { + "epoch": 1.2884667361674196, + "grad_norm": 1.3179954290390015, + "learning_rate": 0.00011410479169079889, + "loss": 0.1794, + "step": 33370 + }, + { + "epoch": 1.2888528514614463, + "grad_norm": 1.0040466785430908, + "learning_rate": 0.00011407905067119709, + "loss": 0.3037, + "step": 33380 + }, + { + "epoch": 1.2892389667554731, + "grad_norm": 5.296288013458252, + "learning_rate": 0.00011405330965159531, + "loss": 0.2904, + "step": 33390 + }, + { + "epoch": 1.2896250820495, + "grad_norm": 0.4267273247241974, + "learning_rate": 0.00011402756863199351, + "loss": 0.2263, + "step": 33400 + }, + { + "epoch": 1.2900111973435266, + "grad_norm": 0.8817713260650635, + "learning_rate": 0.00011400182761239173, + "loss": 0.2715, + "step": 33410 + }, + { + "epoch": 1.2903973126375536, + "grad_norm": 2.7891275882720947, + "learning_rate": 0.00011397608659278995, + "loss": 0.2781, + "step": 33420 + }, + { + "epoch": 1.2907834279315804, + "grad_norm": 0.3013952374458313, + "learning_rate": 0.00011395034557318815, + "loss": 0.2563, + "step": 33430 + }, + { + "epoch": 1.2911695432256072, + "grad_norm": 1.766413927078247, + "learning_rate": 0.00011392460455358638, + "loss": 0.1854, + "step": 33440 + }, + { + "epoch": 1.291555658519634, + "grad_norm": 0.25331103801727295, + "learning_rate": 0.00011389886353398458, + "loss": 0.1206, + "step": 33450 + }, + { + "epoch": 1.2919417738136607, + "grad_norm": 0.036400288343429565, + "learning_rate": 0.0001138731225143828, + "loss": 0.4707, + "step": 33460 + }, + { + "epoch": 1.2923278891076875, + "grad_norm": 1.5620888471603394, + "learning_rate": 0.00011384738149478102, + "loss": 0.3312, + "step": 33470 + }, + { + "epoch": 1.2927140044017142, + "grad_norm": 0.6670392155647278, + "learning_rate": 0.00011382164047517922, + "loss": 0.2341, + "step": 33480 + }, + { + "epoch": 1.2931001196957412, + "grad_norm": 2.3108737468719482, + "learning_rate": 0.00011379589945557745, + "loss": 0.3843, + "step": 33490 + }, + { + "epoch": 1.293486234989768, + "grad_norm": 0.8025147318840027, + "learning_rate": 0.00011377015843597565, + "loss": 0.1982, + "step": 33500 + }, + { + "epoch": 1.2938723502837948, + "grad_norm": 1.7835719585418701, + "learning_rate": 0.00011374441741637387, + "loss": 0.3285, + "step": 33510 + }, + { + "epoch": 1.2942584655778215, + "grad_norm": 2.041508913040161, + "learning_rate": 0.00011371867639677207, + "loss": 0.2044, + "step": 33520 + }, + { + "epoch": 1.2946445808718483, + "grad_norm": 1.103378415107727, + "learning_rate": 0.0001136929353771703, + "loss": 0.1682, + "step": 33530 + }, + { + "epoch": 1.295030696165875, + "grad_norm": 0.057376351207494736, + "learning_rate": 0.00011366719435756851, + "loss": 0.1642, + "step": 33540 + }, + { + "epoch": 1.2954168114599018, + "grad_norm": 0.6539410948753357, + "learning_rate": 0.00011364145333796671, + "loss": 0.1549, + "step": 33550 + }, + { + "epoch": 1.2958029267539288, + "grad_norm": 1.250543236732483, + "learning_rate": 0.00011361571231836494, + "loss": 0.3764, + "step": 33560 + }, + { + "epoch": 1.2961890420479556, + "grad_norm": 0.23697887361049652, + "learning_rate": 0.00011358997129876314, + "loss": 0.3999, + "step": 33570 + }, + { + "epoch": 1.2965751573419824, + "grad_norm": 0.9318505525588989, + "learning_rate": 0.00011356423027916137, + "loss": 0.4156, + "step": 33580 + }, + { + "epoch": 1.2969612726360091, + "grad_norm": 1.3910777568817139, + "learning_rate": 0.00011353848925955957, + "loss": 0.3455, + "step": 33590 + }, + { + "epoch": 1.297347387930036, + "grad_norm": 1.6764451265335083, + "learning_rate": 0.00011351274823995779, + "loss": 0.1884, + "step": 33600 + }, + { + "epoch": 1.2977335032240627, + "grad_norm": 0.9300051927566528, + "learning_rate": 0.000113487007220356, + "loss": 0.119, + "step": 33610 + }, + { + "epoch": 1.2981196185180894, + "grad_norm": 2.447462558746338, + "learning_rate": 0.00011346126620075422, + "loss": 0.4403, + "step": 33620 + }, + { + "epoch": 1.2985057338121164, + "grad_norm": 1.216407060623169, + "learning_rate": 0.00011343552518115243, + "loss": 0.2415, + "step": 33630 + }, + { + "epoch": 1.298891849106143, + "grad_norm": 2.968648910522461, + "learning_rate": 0.00011340978416155063, + "loss": 0.2899, + "step": 33640 + }, + { + "epoch": 1.29927796440017, + "grad_norm": 0.6649970412254333, + "learning_rate": 0.00011338404314194886, + "loss": 0.3809, + "step": 33650 + }, + { + "epoch": 1.2996640796941967, + "grad_norm": 1.7277917861938477, + "learning_rate": 0.00011335830212234709, + "loss": 0.3308, + "step": 33660 + }, + { + "epoch": 1.3000501949882235, + "grad_norm": 1.3269709348678589, + "learning_rate": 0.00011333256110274529, + "loss": 0.3682, + "step": 33670 + }, + { + "epoch": 1.3004363102822503, + "grad_norm": 0.20609407126903534, + "learning_rate": 0.0001133068200831435, + "loss": 0.1379, + "step": 33680 + }, + { + "epoch": 1.300822425576277, + "grad_norm": 0.6592215299606323, + "learning_rate": 0.00011328107906354171, + "loss": 0.2746, + "step": 33690 + }, + { + "epoch": 1.3012085408703038, + "grad_norm": 1.903635859489441, + "learning_rate": 0.00011325533804393993, + "loss": 0.4729, + "step": 33700 + }, + { + "epoch": 1.3015946561643306, + "grad_norm": 0.8432504534721375, + "learning_rate": 0.00011322959702433813, + "loss": 0.2835, + "step": 33710 + }, + { + "epoch": 1.3019807714583576, + "grad_norm": 0.9862542152404785, + "learning_rate": 0.00011320385600473635, + "loss": 0.1823, + "step": 33720 + }, + { + "epoch": 1.3023668867523843, + "grad_norm": 3.845738649368286, + "learning_rate": 0.00011317811498513458, + "loss": 0.2542, + "step": 33730 + }, + { + "epoch": 1.302753002046411, + "grad_norm": 0.6317747235298157, + "learning_rate": 0.00011315237396553278, + "loss": 0.22, + "step": 33740 + }, + { + "epoch": 1.3031391173404379, + "grad_norm": 2.5221354961395264, + "learning_rate": 0.000113126632945931, + "loss": 0.2253, + "step": 33750 + }, + { + "epoch": 1.3035252326344646, + "grad_norm": 1.3326247930526733, + "learning_rate": 0.0001131008919263292, + "loss": 0.2021, + "step": 33760 + }, + { + "epoch": 1.3039113479284914, + "grad_norm": 1.148047685623169, + "learning_rate": 0.00011307515090672742, + "loss": 0.3987, + "step": 33770 + }, + { + "epoch": 1.3042974632225182, + "grad_norm": 0.19721268117427826, + "learning_rate": 0.00011304940988712562, + "loss": 0.2642, + "step": 33780 + }, + { + "epoch": 1.3046835785165452, + "grad_norm": 1.4060617685317993, + "learning_rate": 0.00011302366886752385, + "loss": 0.2736, + "step": 33790 + }, + { + "epoch": 1.305069693810572, + "grad_norm": 1.0736548900604248, + "learning_rate": 0.00011299792784792207, + "loss": 0.2448, + "step": 33800 + }, + { + "epoch": 1.3054558091045987, + "grad_norm": 4.352476119995117, + "learning_rate": 0.00011297218682832027, + "loss": 0.383, + "step": 33810 + }, + { + "epoch": 1.3058419243986255, + "grad_norm": 0.2249228060245514, + "learning_rate": 0.0001129464458087185, + "loss": 0.14, + "step": 33820 + }, + { + "epoch": 1.3062280396926522, + "grad_norm": 0.4820781946182251, + "learning_rate": 0.0001129207047891167, + "loss": 0.248, + "step": 33830 + }, + { + "epoch": 1.306614154986679, + "grad_norm": 2.2983391284942627, + "learning_rate": 0.00011289496376951491, + "loss": 0.2608, + "step": 33840 + }, + { + "epoch": 1.3070002702807058, + "grad_norm": 1.3315671682357788, + "learning_rate": 0.00011286922274991314, + "loss": 0.1727, + "step": 33850 + }, + { + "epoch": 1.3073863855747327, + "grad_norm": 2.060299873352051, + "learning_rate": 0.00011284348173031134, + "loss": 0.3097, + "step": 33860 + }, + { + "epoch": 1.3077725008687593, + "grad_norm": 2.096285581588745, + "learning_rate": 0.00011281774071070957, + "loss": 0.2305, + "step": 33870 + }, + { + "epoch": 1.3081586161627863, + "grad_norm": 0.4997636675834656, + "learning_rate": 0.00011279199969110777, + "loss": 0.1993, + "step": 33880 + }, + { + "epoch": 1.308544731456813, + "grad_norm": 0.58636474609375, + "learning_rate": 0.00011276625867150599, + "loss": 0.1873, + "step": 33890 + }, + { + "epoch": 1.3089308467508398, + "grad_norm": 0.9128592610359192, + "learning_rate": 0.00011274051765190419, + "loss": 0.1885, + "step": 33900 + }, + { + "epoch": 1.3093169620448666, + "grad_norm": 2.228043794631958, + "learning_rate": 0.0001127147766323024, + "loss": 0.3649, + "step": 33910 + }, + { + "epoch": 1.3097030773388934, + "grad_norm": 1.069002389907837, + "learning_rate": 0.00011268903561270063, + "loss": 0.5454, + "step": 33920 + }, + { + "epoch": 1.3100891926329201, + "grad_norm": 0.6207597851753235, + "learning_rate": 0.00011266329459309883, + "loss": 0.2329, + "step": 33930 + }, + { + "epoch": 1.3104753079269469, + "grad_norm": 1.262247920036316, + "learning_rate": 0.00011263755357349706, + "loss": 0.3437, + "step": 33940 + }, + { + "epoch": 1.3108614232209739, + "grad_norm": 1.7429994344711304, + "learning_rate": 0.00011261181255389526, + "loss": 0.228, + "step": 33950 + }, + { + "epoch": 1.3112475385150006, + "grad_norm": 0.646900475025177, + "learning_rate": 0.00011258607153429349, + "loss": 0.3739, + "step": 33960 + }, + { + "epoch": 1.3116336538090274, + "grad_norm": 1.8228782415390015, + "learning_rate": 0.0001125603305146917, + "loss": 0.2325, + "step": 33970 + }, + { + "epoch": 1.3120197691030542, + "grad_norm": 3.539228916168213, + "learning_rate": 0.0001125345894950899, + "loss": 0.194, + "step": 33980 + }, + { + "epoch": 1.312405884397081, + "grad_norm": 1.2801135778427124, + "learning_rate": 0.00011250884847548813, + "loss": 0.3069, + "step": 33990 + }, + { + "epoch": 1.3127919996911077, + "grad_norm": 3.6265695095062256, + "learning_rate": 0.00011248310745588633, + "loss": 0.3113, + "step": 34000 + }, + { + "epoch": 1.3131781149851345, + "grad_norm": 0.07370063662528992, + "learning_rate": 0.00011245736643628455, + "loss": 0.1449, + "step": 34010 + }, + { + "epoch": 1.3135642302791615, + "grad_norm": 1.0295637845993042, + "learning_rate": 0.00011243162541668275, + "loss": 0.22, + "step": 34020 + }, + { + "epoch": 1.313950345573188, + "grad_norm": 0.8803662061691284, + "learning_rate": 0.00011240588439708098, + "loss": 0.1368, + "step": 34030 + }, + { + "epoch": 1.314336460867215, + "grad_norm": 1.6597707271575928, + "learning_rate": 0.00011238014337747919, + "loss": 0.3038, + "step": 34040 + }, + { + "epoch": 1.3147225761612418, + "grad_norm": 2.115492343902588, + "learning_rate": 0.00011235440235787739, + "loss": 0.1754, + "step": 34050 + }, + { + "epoch": 1.3151086914552685, + "grad_norm": 0.8143919706344604, + "learning_rate": 0.00011232866133827562, + "loss": 0.3764, + "step": 34060 + }, + { + "epoch": 1.3154948067492953, + "grad_norm": 0.14369767904281616, + "learning_rate": 0.00011230292031867382, + "loss": 0.1742, + "step": 34070 + }, + { + "epoch": 1.315880922043322, + "grad_norm": 1.0129845142364502, + "learning_rate": 0.00011227717929907205, + "loss": 0.1458, + "step": 34080 + }, + { + "epoch": 1.316267037337349, + "grad_norm": 2.7300291061401367, + "learning_rate": 0.00011225143827947025, + "loss": 0.3939, + "step": 34090 + }, + { + "epoch": 1.3166531526313756, + "grad_norm": 0.20205609500408173, + "learning_rate": 0.00011222569725986847, + "loss": 0.29, + "step": 34100 + }, + { + "epoch": 1.3170392679254026, + "grad_norm": 1.8928464651107788, + "learning_rate": 0.00011219995624026669, + "loss": 0.1742, + "step": 34110 + }, + { + "epoch": 1.3174253832194294, + "grad_norm": 0.2639687955379486, + "learning_rate": 0.00011217421522066488, + "loss": 0.1745, + "step": 34120 + }, + { + "epoch": 1.3178114985134561, + "grad_norm": 0.5906389355659485, + "learning_rate": 0.00011214847420106311, + "loss": 0.2134, + "step": 34130 + }, + { + "epoch": 1.318197613807483, + "grad_norm": 0.9190629720687866, + "learning_rate": 0.00011212273318146131, + "loss": 0.2547, + "step": 34140 + }, + { + "epoch": 1.3185837291015097, + "grad_norm": 0.5749151110649109, + "learning_rate": 0.00011209699216185954, + "loss": 0.1688, + "step": 34150 + }, + { + "epoch": 1.3189698443955364, + "grad_norm": 0.82295823097229, + "learning_rate": 0.00011207125114225777, + "loss": 0.2884, + "step": 34160 + }, + { + "epoch": 1.3193559596895632, + "grad_norm": 0.07816460728645325, + "learning_rate": 0.00011204551012265597, + "loss": 0.2418, + "step": 34170 + }, + { + "epoch": 1.3197420749835902, + "grad_norm": 0.6417407393455505, + "learning_rate": 0.00011201976910305418, + "loss": 0.2557, + "step": 34180 + }, + { + "epoch": 1.320128190277617, + "grad_norm": 6.093267440795898, + "learning_rate": 0.00011199402808345238, + "loss": 0.3088, + "step": 34190 + }, + { + "epoch": 1.3205143055716437, + "grad_norm": 1.8861887454986572, + "learning_rate": 0.0001119682870638506, + "loss": 0.2204, + "step": 34200 + }, + { + "epoch": 1.3209004208656705, + "grad_norm": 2.3272714614868164, + "learning_rate": 0.0001119425460442488, + "loss": 0.2236, + "step": 34210 + }, + { + "epoch": 1.3212865361596973, + "grad_norm": 0.9608810544013977, + "learning_rate": 0.00011191680502464703, + "loss": 0.1897, + "step": 34220 + }, + { + "epoch": 1.321672651453724, + "grad_norm": 1.2157350778579712, + "learning_rate": 0.00011189106400504526, + "loss": 0.1526, + "step": 34230 + }, + { + "epoch": 1.3220587667477508, + "grad_norm": 1.6684671640396118, + "learning_rate": 0.00011186532298544346, + "loss": 0.3394, + "step": 34240 + }, + { + "epoch": 1.3224448820417778, + "grad_norm": 2.0432374477386475, + "learning_rate": 0.00011183958196584167, + "loss": 0.2183, + "step": 34250 + }, + { + "epoch": 1.3228309973358043, + "grad_norm": 0.9436892867088318, + "learning_rate": 0.00011181384094623988, + "loss": 0.2947, + "step": 34260 + }, + { + "epoch": 1.3232171126298313, + "grad_norm": 0.23260092735290527, + "learning_rate": 0.0001117880999266381, + "loss": 0.114, + "step": 34270 + }, + { + "epoch": 1.323603227923858, + "grad_norm": 1.2291594743728638, + "learning_rate": 0.0001117623589070363, + "loss": 0.3145, + "step": 34280 + }, + { + "epoch": 1.3239893432178849, + "grad_norm": 0.41411107778549194, + "learning_rate": 0.00011173661788743452, + "loss": 0.2937, + "step": 34290 + }, + { + "epoch": 1.3243754585119116, + "grad_norm": 2.354405164718628, + "learning_rate": 0.00011171087686783275, + "loss": 0.3933, + "step": 34300 + }, + { + "epoch": 1.3247615738059384, + "grad_norm": 2.6997978687286377, + "learning_rate": 0.00011168513584823095, + "loss": 0.1494, + "step": 34310 + }, + { + "epoch": 1.3251476890999654, + "grad_norm": 2.8430919647216797, + "learning_rate": 0.00011165939482862916, + "loss": 0.2869, + "step": 34320 + }, + { + "epoch": 1.325533804393992, + "grad_norm": 1.1737356185913086, + "learning_rate": 0.00011163365380902738, + "loss": 0.2792, + "step": 34330 + }, + { + "epoch": 1.325919919688019, + "grad_norm": 4.123973846435547, + "learning_rate": 0.00011160791278942559, + "loss": 0.5211, + "step": 34340 + }, + { + "epoch": 1.3263060349820457, + "grad_norm": 0.8862038850784302, + "learning_rate": 0.00011158217176982382, + "loss": 0.2976, + "step": 34350 + }, + { + "epoch": 1.3266921502760725, + "grad_norm": 1.8690590858459473, + "learning_rate": 0.00011155643075022202, + "loss": 0.2485, + "step": 34360 + }, + { + "epoch": 1.3270782655700992, + "grad_norm": 2.885589599609375, + "learning_rate": 0.00011153068973062024, + "loss": 0.3084, + "step": 34370 + }, + { + "epoch": 1.327464380864126, + "grad_norm": 0.9898788928985596, + "learning_rate": 0.00011150494871101844, + "loss": 0.261, + "step": 34380 + }, + { + "epoch": 1.3278504961581528, + "grad_norm": 0.6879653930664062, + "learning_rate": 0.00011147920769141667, + "loss": 0.2082, + "step": 34390 + }, + { + "epoch": 1.3282366114521795, + "grad_norm": 1.2619003057479858, + "learning_rate": 0.00011145346667181487, + "loss": 0.2402, + "step": 34400 + }, + { + "epoch": 1.3286227267462065, + "grad_norm": 1.1212007999420166, + "learning_rate": 0.00011142772565221308, + "loss": 0.3525, + "step": 34410 + }, + { + "epoch": 1.3290088420402333, + "grad_norm": 1.8431956768035889, + "learning_rate": 0.00011140198463261131, + "loss": 0.2212, + "step": 34420 + }, + { + "epoch": 1.32939495733426, + "grad_norm": 0.6185423731803894, + "learning_rate": 0.00011137624361300951, + "loss": 0.2455, + "step": 34430 + }, + { + "epoch": 1.3297810726282868, + "grad_norm": 2.3791301250457764, + "learning_rate": 0.00011135050259340774, + "loss": 0.1763, + "step": 34440 + }, + { + "epoch": 1.3301671879223136, + "grad_norm": 0.4928603768348694, + "learning_rate": 0.00011132476157380594, + "loss": 0.2381, + "step": 34450 + }, + { + "epoch": 1.3305533032163404, + "grad_norm": 1.5636029243469238, + "learning_rate": 0.00011129902055420416, + "loss": 0.1368, + "step": 34460 + }, + { + "epoch": 1.3309394185103671, + "grad_norm": 0.9425283074378967, + "learning_rate": 0.00011127327953460238, + "loss": 0.2825, + "step": 34470 + }, + { + "epoch": 1.3313255338043941, + "grad_norm": 1.2257115840911865, + "learning_rate": 0.00011124753851500058, + "loss": 0.2547, + "step": 34480 + }, + { + "epoch": 1.3317116490984207, + "grad_norm": 0.9416170716285706, + "learning_rate": 0.0001112217974953988, + "loss": 0.2766, + "step": 34490 + }, + { + "epoch": 1.3320977643924476, + "grad_norm": 0.5123847126960754, + "learning_rate": 0.000111196056475797, + "loss": 0.4733, + "step": 34500 + }, + { + "epoch": 1.3324838796864744, + "grad_norm": 1.5581384897232056, + "learning_rate": 0.00011117031545619523, + "loss": 0.1597, + "step": 34510 + }, + { + "epoch": 1.3328699949805012, + "grad_norm": 2.377333879470825, + "learning_rate": 0.00011114457443659343, + "loss": 0.209, + "step": 34520 + }, + { + "epoch": 1.333256110274528, + "grad_norm": 1.7840913534164429, + "learning_rate": 0.00011111883341699166, + "loss": 0.1759, + "step": 34530 + }, + { + "epoch": 1.3336422255685547, + "grad_norm": 1.1825993061065674, + "learning_rate": 0.00011109309239738987, + "loss": 0.2464, + "step": 34540 + }, + { + "epoch": 1.3340283408625815, + "grad_norm": 1.8859659433364868, + "learning_rate": 0.00011106735137778807, + "loss": 0.3539, + "step": 34550 + }, + { + "epoch": 1.3344144561566083, + "grad_norm": 1.9698175191879272, + "learning_rate": 0.0001110416103581863, + "loss": 0.3301, + "step": 34560 + }, + { + "epoch": 1.3348005714506352, + "grad_norm": 0.7649385333061218, + "learning_rate": 0.0001110158693385845, + "loss": 0.232, + "step": 34570 + }, + { + "epoch": 1.335186686744662, + "grad_norm": 0.56386399269104, + "learning_rate": 0.00011099012831898272, + "loss": 0.3425, + "step": 34580 + }, + { + "epoch": 1.3355728020386888, + "grad_norm": 2.956003189086914, + "learning_rate": 0.00011096438729938092, + "loss": 0.1518, + "step": 34590 + }, + { + "epoch": 1.3359589173327155, + "grad_norm": 2.612029552459717, + "learning_rate": 0.00011093864627977915, + "loss": 0.2765, + "step": 34600 + }, + { + "epoch": 1.3363450326267423, + "grad_norm": 0.9674397706985474, + "learning_rate": 0.00011091290526017736, + "loss": 0.303, + "step": 34610 + }, + { + "epoch": 1.336731147920769, + "grad_norm": 0.9578921794891357, + "learning_rate": 0.00011088716424057556, + "loss": 0.1405, + "step": 34620 + }, + { + "epoch": 1.3371172632147958, + "grad_norm": 2.168065071105957, + "learning_rate": 0.00011086142322097379, + "loss": 0.1914, + "step": 34630 + }, + { + "epoch": 1.3375033785088228, + "grad_norm": 1.3166526556015015, + "learning_rate": 0.00011083568220137199, + "loss": 0.4134, + "step": 34640 + }, + { + "epoch": 1.3378894938028496, + "grad_norm": 0.9082283973693848, + "learning_rate": 0.00011080994118177022, + "loss": 0.2693, + "step": 34650 + }, + { + "epoch": 1.3382756090968764, + "grad_norm": 2.203007698059082, + "learning_rate": 0.00011078420016216844, + "loss": 0.1847, + "step": 34660 + }, + { + "epoch": 1.3386617243909031, + "grad_norm": 0.8101674914360046, + "learning_rate": 0.00011075845914256664, + "loss": 0.3111, + "step": 34670 + }, + { + "epoch": 1.33904783968493, + "grad_norm": 1.9545695781707764, + "learning_rate": 0.00011073271812296486, + "loss": 0.3761, + "step": 34680 + }, + { + "epoch": 1.3394339549789567, + "grad_norm": 1.547581672668457, + "learning_rate": 0.00011070697710336306, + "loss": 0.2374, + "step": 34690 + }, + { + "epoch": 1.3398200702729834, + "grad_norm": 3.3519034385681152, + "learning_rate": 0.00011068123608376128, + "loss": 0.1957, + "step": 34700 + }, + { + "epoch": 1.3402061855670104, + "grad_norm": 1.5508599281311035, + "learning_rate": 0.00011065549506415948, + "loss": 0.4171, + "step": 34710 + }, + { + "epoch": 1.340592300861037, + "grad_norm": 1.8547546863555908, + "learning_rate": 0.00011062975404455771, + "loss": 0.1872, + "step": 34720 + }, + { + "epoch": 1.340978416155064, + "grad_norm": 1.4600756168365479, + "learning_rate": 0.00011060401302495594, + "loss": 0.3515, + "step": 34730 + }, + { + "epoch": 1.3413645314490907, + "grad_norm": 0.05774044618010521, + "learning_rate": 0.00011057827200535414, + "loss": 0.1604, + "step": 34740 + }, + { + "epoch": 1.3417506467431175, + "grad_norm": 2.8793342113494873, + "learning_rate": 0.00011055253098575235, + "loss": 0.3095, + "step": 34750 + }, + { + "epoch": 1.3421367620371443, + "grad_norm": 2.241042375564575, + "learning_rate": 0.00011052678996615055, + "loss": 0.2511, + "step": 34760 + }, + { + "epoch": 1.342522877331171, + "grad_norm": 1.9320632219314575, + "learning_rate": 0.00011050104894654878, + "loss": 0.4493, + "step": 34770 + }, + { + "epoch": 1.3429089926251978, + "grad_norm": 1.6483882665634155, + "learning_rate": 0.000110475307926947, + "loss": 0.217, + "step": 34780 + }, + { + "epoch": 1.3432951079192246, + "grad_norm": 0.9635765552520752, + "learning_rate": 0.0001104495669073452, + "loss": 0.5458, + "step": 34790 + }, + { + "epoch": 1.3436812232132516, + "grad_norm": 1.2436567544937134, + "learning_rate": 0.00011042382588774343, + "loss": 0.2857, + "step": 34800 + }, + { + "epoch": 1.3440673385072783, + "grad_norm": 2.8082425594329834, + "learning_rate": 0.00011039808486814163, + "loss": 0.3439, + "step": 34810 + }, + { + "epoch": 1.344453453801305, + "grad_norm": 1.0430901050567627, + "learning_rate": 0.00011037234384853984, + "loss": 0.1404, + "step": 34820 + }, + { + "epoch": 1.3448395690953319, + "grad_norm": 1.7387149333953857, + "learning_rate": 0.00011034660282893806, + "loss": 0.395, + "step": 34830 + }, + { + "epoch": 1.3452256843893586, + "grad_norm": 1.2713748216629028, + "learning_rate": 0.00011032086180933627, + "loss": 0.29, + "step": 34840 + }, + { + "epoch": 1.3456117996833854, + "grad_norm": 0.26068204641342163, + "learning_rate": 0.0001102951207897345, + "loss": 0.1814, + "step": 34850 + }, + { + "epoch": 1.3459979149774122, + "grad_norm": 2.4163243770599365, + "learning_rate": 0.0001102693797701327, + "loss": 0.2445, + "step": 34860 + }, + { + "epoch": 1.3463840302714392, + "grad_norm": 2.2439687252044678, + "learning_rate": 0.00011024363875053092, + "loss": 0.3338, + "step": 34870 + }, + { + "epoch": 1.346770145565466, + "grad_norm": 0.2822403609752655, + "learning_rate": 0.00011021789773092912, + "loss": 0.1648, + "step": 34880 + }, + { + "epoch": 1.3471562608594927, + "grad_norm": 0.07319017499685287, + "learning_rate": 0.00011019215671132734, + "loss": 0.107, + "step": 34890 + }, + { + "epoch": 1.3475423761535195, + "grad_norm": 0.9809044003486633, + "learning_rate": 0.00011016641569172555, + "loss": 0.256, + "step": 34900 + }, + { + "epoch": 1.3479284914475462, + "grad_norm": 0.5016226768493652, + "learning_rate": 0.00011014067467212376, + "loss": 0.3229, + "step": 34910 + }, + { + "epoch": 1.348314606741573, + "grad_norm": 1.3026005029678345, + "learning_rate": 0.00011011493365252199, + "loss": 0.2145, + "step": 34920 + }, + { + "epoch": 1.3487007220355998, + "grad_norm": 1.0752215385437012, + "learning_rate": 0.00011008919263292019, + "loss": 0.2355, + "step": 34930 + }, + { + "epoch": 1.3490868373296268, + "grad_norm": 2.2703003883361816, + "learning_rate": 0.00011006345161331842, + "loss": 0.2079, + "step": 34940 + }, + { + "epoch": 1.3494729526236533, + "grad_norm": 1.1323810815811157, + "learning_rate": 0.00011003771059371662, + "loss": 0.1015, + "step": 34950 + }, + { + "epoch": 1.3498590679176803, + "grad_norm": 0.10813555121421814, + "learning_rate": 0.00011001196957411484, + "loss": 0.4214, + "step": 34960 + }, + { + "epoch": 1.350245183211707, + "grad_norm": 0.07815568149089813, + "learning_rate": 0.00010998622855451306, + "loss": 0.1211, + "step": 34970 + }, + { + "epoch": 1.3506312985057338, + "grad_norm": 0.6748234629631042, + "learning_rate": 0.00010996048753491126, + "loss": 0.3508, + "step": 34980 + }, + { + "epoch": 1.3510174137997606, + "grad_norm": 1.8556997776031494, + "learning_rate": 0.00010993474651530948, + "loss": 0.2268, + "step": 34990 + }, + { + "epoch": 1.3514035290937874, + "grad_norm": 0.8696061372756958, + "learning_rate": 0.00010990900549570768, + "loss": 0.4321, + "step": 35000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.13135261696e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f6975dee08d9cd543f202792d4d9c34c7f6ca861 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55d1af15113482c6842dfa4a381017ad95d55077f598c749d5d0379d930a7843 +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..ba2b5397c1c7ee71fa1654f58ad7564bc3adaeec --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1da79ec115521040017d1b194500aa1eaa28afbd599770f2443cb48e8f8ea9b +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..14298cf3a9b43eb3e3e884ae07398ddd0070ca89 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..86fbd86230c9660af8d090656af3b6a375b41a9c Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..1d7b6a5c750f5a5d32a9947c371927590aacbc66 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/trainer_state.json @@ -0,0 +1,24884 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.3707092937951273, + "eval_steps": 500, + "global_step": 35500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + }, + { + "epoch": 1.274566585582455, + "grad_norm": 0.5019121766090393, + "learning_rate": 0.0001150314683964632, + "loss": 0.1857, + "step": 33010 + }, + { + "epoch": 1.2749527008764816, + "grad_norm": 0.258952260017395, + "learning_rate": 0.0001150057273768614, + "loss": 0.11, + "step": 33020 + }, + { + "epoch": 1.2753388161705086, + "grad_norm": 0.8540348410606384, + "learning_rate": 0.00011497998635725962, + "loss": 0.1852, + "step": 33030 + }, + { + "epoch": 1.2757249314645354, + "grad_norm": 0.08689398318529129, + "learning_rate": 0.00011495424533765783, + "loss": 0.2225, + "step": 33040 + }, + { + "epoch": 1.2761110467585621, + "grad_norm": 0.42253822088241577, + "learning_rate": 0.00011492850431805605, + "loss": 0.0751, + "step": 33050 + }, + { + "epoch": 1.276497162052589, + "grad_norm": 1.2964017391204834, + "learning_rate": 0.00011490276329845426, + "loss": 0.2384, + "step": 33060 + }, + { + "epoch": 1.2768832773466157, + "grad_norm": 0.5337836146354675, + "learning_rate": 0.00011487702227885246, + "loss": 0.1415, + "step": 33070 + }, + { + "epoch": 1.2772693926406424, + "grad_norm": 2.7771682739257812, + "learning_rate": 0.00011485128125925069, + "loss": 0.378, + "step": 33080 + }, + { + "epoch": 1.2776555079346692, + "grad_norm": 1.5107232332229614, + "learning_rate": 0.00011482554023964889, + "loss": 0.2482, + "step": 33090 + }, + { + "epoch": 1.2780416232286962, + "grad_norm": 0.6885499358177185, + "learning_rate": 0.00011479979922004711, + "loss": 0.2122, + "step": 33100 + }, + { + "epoch": 1.278427738522723, + "grad_norm": 0.9016557335853577, + "learning_rate": 0.00011477405820044533, + "loss": 0.2841, + "step": 33110 + }, + { + "epoch": 1.2788138538167497, + "grad_norm": 1.9532525539398193, + "learning_rate": 0.00011474831718084354, + "loss": 0.2281, + "step": 33120 + }, + { + "epoch": 1.2791999691107765, + "grad_norm": 2.1078782081604004, + "learning_rate": 0.00011472257616124175, + "loss": 0.2836, + "step": 33130 + }, + { + "epoch": 1.2795860844048033, + "grad_norm": 0.19830390810966492, + "learning_rate": 0.00011469683514163995, + "loss": 0.388, + "step": 33140 + }, + { + "epoch": 1.27997219969883, + "grad_norm": 0.17538850009441376, + "learning_rate": 0.00011467109412203818, + "loss": 0.3274, + "step": 33150 + }, + { + "epoch": 1.2803583149928568, + "grad_norm": 0.7402139902114868, + "learning_rate": 0.00011464535310243641, + "loss": 0.1979, + "step": 33160 + }, + { + "epoch": 1.2807444302868838, + "grad_norm": 0.2097146362066269, + "learning_rate": 0.00011461961208283461, + "loss": 0.2464, + "step": 33170 + }, + { + "epoch": 1.2811305455809103, + "grad_norm": 1.2441083192825317, + "learning_rate": 0.00011459387106323283, + "loss": 0.266, + "step": 33180 + }, + { + "epoch": 1.2815166608749373, + "grad_norm": 2.518852710723877, + "learning_rate": 0.00011456813004363103, + "loss": 0.253, + "step": 33190 + }, + { + "epoch": 1.281902776168964, + "grad_norm": 0.8078998327255249, + "learning_rate": 0.00011454238902402925, + "loss": 0.2361, + "step": 33200 + }, + { + "epoch": 1.2822888914629909, + "grad_norm": 1.2297371625900269, + "learning_rate": 0.00011451664800442745, + "loss": 0.1974, + "step": 33210 + }, + { + "epoch": 1.2826750067570176, + "grad_norm": 0.4303855895996094, + "learning_rate": 0.00011449090698482567, + "loss": 0.3563, + "step": 33220 + }, + { + "epoch": 1.2830611220510444, + "grad_norm": 1.3215210437774658, + "learning_rate": 0.0001144651659652239, + "loss": 0.2818, + "step": 33230 + }, + { + "epoch": 1.2834472373450712, + "grad_norm": 1.546265959739685, + "learning_rate": 0.0001144394249456221, + "loss": 0.5778, + "step": 33240 + }, + { + "epoch": 1.283833352639098, + "grad_norm": 0.8895953297615051, + "learning_rate": 0.00011441368392602033, + "loss": 0.2485, + "step": 33250 + }, + { + "epoch": 1.284219467933125, + "grad_norm": 0.7534870505332947, + "learning_rate": 0.00011438794290641853, + "loss": 0.2825, + "step": 33260 + }, + { + "epoch": 1.2846055832271517, + "grad_norm": 0.052820973098278046, + "learning_rate": 0.00011436220188681674, + "loss": 0.2191, + "step": 33270 + }, + { + "epoch": 1.2849916985211784, + "grad_norm": 0.9264475107192993, + "learning_rate": 0.00011433646086721494, + "loss": 0.181, + "step": 33280 + }, + { + "epoch": 1.2853778138152052, + "grad_norm": 0.2128441333770752, + "learning_rate": 0.00011431071984761317, + "loss": 0.1819, + "step": 33290 + }, + { + "epoch": 1.285763929109232, + "grad_norm": 0.5400950312614441, + "learning_rate": 0.0001142849788280114, + "loss": 0.4316, + "step": 33300 + }, + { + "epoch": 1.2861500444032588, + "grad_norm": 0.8033271431922913, + "learning_rate": 0.00011425923780840959, + "loss": 0.2146, + "step": 33310 + }, + { + "epoch": 1.2865361596972855, + "grad_norm": 2.012575149536133, + "learning_rate": 0.00011423349678880782, + "loss": 0.4335, + "step": 33320 + }, + { + "epoch": 1.2869222749913125, + "grad_norm": 0.7352376580238342, + "learning_rate": 0.00011420775576920602, + "loss": 0.2124, + "step": 33330 + }, + { + "epoch": 1.2873083902853393, + "grad_norm": 0.769036591053009, + "learning_rate": 0.00011418201474960423, + "loss": 0.3602, + "step": 33340 + }, + { + "epoch": 1.287694505579366, + "grad_norm": 0.250592976808548, + "learning_rate": 0.00011415627373000246, + "loss": 0.1692, + "step": 33350 + }, + { + "epoch": 1.2880806208733928, + "grad_norm": 2.43820858001709, + "learning_rate": 0.00011413053271040066, + "loss": 0.2777, + "step": 33360 + }, + { + "epoch": 1.2884667361674196, + "grad_norm": 1.3179954290390015, + "learning_rate": 0.00011410479169079889, + "loss": 0.1794, + "step": 33370 + }, + { + "epoch": 1.2888528514614463, + "grad_norm": 1.0040466785430908, + "learning_rate": 0.00011407905067119709, + "loss": 0.3037, + "step": 33380 + }, + { + "epoch": 1.2892389667554731, + "grad_norm": 5.296288013458252, + "learning_rate": 0.00011405330965159531, + "loss": 0.2904, + "step": 33390 + }, + { + "epoch": 1.2896250820495, + "grad_norm": 0.4267273247241974, + "learning_rate": 0.00011402756863199351, + "loss": 0.2263, + "step": 33400 + }, + { + "epoch": 1.2900111973435266, + "grad_norm": 0.8817713260650635, + "learning_rate": 0.00011400182761239173, + "loss": 0.2715, + "step": 33410 + }, + { + "epoch": 1.2903973126375536, + "grad_norm": 2.7891275882720947, + "learning_rate": 0.00011397608659278995, + "loss": 0.2781, + "step": 33420 + }, + { + "epoch": 1.2907834279315804, + "grad_norm": 0.3013952374458313, + "learning_rate": 0.00011395034557318815, + "loss": 0.2563, + "step": 33430 + }, + { + "epoch": 1.2911695432256072, + "grad_norm": 1.766413927078247, + "learning_rate": 0.00011392460455358638, + "loss": 0.1854, + "step": 33440 + }, + { + "epoch": 1.291555658519634, + "grad_norm": 0.25331103801727295, + "learning_rate": 0.00011389886353398458, + "loss": 0.1206, + "step": 33450 + }, + { + "epoch": 1.2919417738136607, + "grad_norm": 0.036400288343429565, + "learning_rate": 0.0001138731225143828, + "loss": 0.4707, + "step": 33460 + }, + { + "epoch": 1.2923278891076875, + "grad_norm": 1.5620888471603394, + "learning_rate": 0.00011384738149478102, + "loss": 0.3312, + "step": 33470 + }, + { + "epoch": 1.2927140044017142, + "grad_norm": 0.6670392155647278, + "learning_rate": 0.00011382164047517922, + "loss": 0.2341, + "step": 33480 + }, + { + "epoch": 1.2931001196957412, + "grad_norm": 2.3108737468719482, + "learning_rate": 0.00011379589945557745, + "loss": 0.3843, + "step": 33490 + }, + { + "epoch": 1.293486234989768, + "grad_norm": 0.8025147318840027, + "learning_rate": 0.00011377015843597565, + "loss": 0.1982, + "step": 33500 + }, + { + "epoch": 1.2938723502837948, + "grad_norm": 1.7835719585418701, + "learning_rate": 0.00011374441741637387, + "loss": 0.3285, + "step": 33510 + }, + { + "epoch": 1.2942584655778215, + "grad_norm": 2.041508913040161, + "learning_rate": 0.00011371867639677207, + "loss": 0.2044, + "step": 33520 + }, + { + "epoch": 1.2946445808718483, + "grad_norm": 1.103378415107727, + "learning_rate": 0.0001136929353771703, + "loss": 0.1682, + "step": 33530 + }, + { + "epoch": 1.295030696165875, + "grad_norm": 0.057376351207494736, + "learning_rate": 0.00011366719435756851, + "loss": 0.1642, + "step": 33540 + }, + { + "epoch": 1.2954168114599018, + "grad_norm": 0.6539410948753357, + "learning_rate": 0.00011364145333796671, + "loss": 0.1549, + "step": 33550 + }, + { + "epoch": 1.2958029267539288, + "grad_norm": 1.250543236732483, + "learning_rate": 0.00011361571231836494, + "loss": 0.3764, + "step": 33560 + }, + { + "epoch": 1.2961890420479556, + "grad_norm": 0.23697887361049652, + "learning_rate": 0.00011358997129876314, + "loss": 0.3999, + "step": 33570 + }, + { + "epoch": 1.2965751573419824, + "grad_norm": 0.9318505525588989, + "learning_rate": 0.00011356423027916137, + "loss": 0.4156, + "step": 33580 + }, + { + "epoch": 1.2969612726360091, + "grad_norm": 1.3910777568817139, + "learning_rate": 0.00011353848925955957, + "loss": 0.3455, + "step": 33590 + }, + { + "epoch": 1.297347387930036, + "grad_norm": 1.6764451265335083, + "learning_rate": 0.00011351274823995779, + "loss": 0.1884, + "step": 33600 + }, + { + "epoch": 1.2977335032240627, + "grad_norm": 0.9300051927566528, + "learning_rate": 0.000113487007220356, + "loss": 0.119, + "step": 33610 + }, + { + "epoch": 1.2981196185180894, + "grad_norm": 2.447462558746338, + "learning_rate": 0.00011346126620075422, + "loss": 0.4403, + "step": 33620 + }, + { + "epoch": 1.2985057338121164, + "grad_norm": 1.216407060623169, + "learning_rate": 0.00011343552518115243, + "loss": 0.2415, + "step": 33630 + }, + { + "epoch": 1.298891849106143, + "grad_norm": 2.968648910522461, + "learning_rate": 0.00011340978416155063, + "loss": 0.2899, + "step": 33640 + }, + { + "epoch": 1.29927796440017, + "grad_norm": 0.6649970412254333, + "learning_rate": 0.00011338404314194886, + "loss": 0.3809, + "step": 33650 + }, + { + "epoch": 1.2996640796941967, + "grad_norm": 1.7277917861938477, + "learning_rate": 0.00011335830212234709, + "loss": 0.3308, + "step": 33660 + }, + { + "epoch": 1.3000501949882235, + "grad_norm": 1.3269709348678589, + "learning_rate": 0.00011333256110274529, + "loss": 0.3682, + "step": 33670 + }, + { + "epoch": 1.3004363102822503, + "grad_norm": 0.20609407126903534, + "learning_rate": 0.0001133068200831435, + "loss": 0.1379, + "step": 33680 + }, + { + "epoch": 1.300822425576277, + "grad_norm": 0.6592215299606323, + "learning_rate": 0.00011328107906354171, + "loss": 0.2746, + "step": 33690 + }, + { + "epoch": 1.3012085408703038, + "grad_norm": 1.903635859489441, + "learning_rate": 0.00011325533804393993, + "loss": 0.4729, + "step": 33700 + }, + { + "epoch": 1.3015946561643306, + "grad_norm": 0.8432504534721375, + "learning_rate": 0.00011322959702433813, + "loss": 0.2835, + "step": 33710 + }, + { + "epoch": 1.3019807714583576, + "grad_norm": 0.9862542152404785, + "learning_rate": 0.00011320385600473635, + "loss": 0.1823, + "step": 33720 + }, + { + "epoch": 1.3023668867523843, + "grad_norm": 3.845738649368286, + "learning_rate": 0.00011317811498513458, + "loss": 0.2542, + "step": 33730 + }, + { + "epoch": 1.302753002046411, + "grad_norm": 0.6317747235298157, + "learning_rate": 0.00011315237396553278, + "loss": 0.22, + "step": 33740 + }, + { + "epoch": 1.3031391173404379, + "grad_norm": 2.5221354961395264, + "learning_rate": 0.000113126632945931, + "loss": 0.2253, + "step": 33750 + }, + { + "epoch": 1.3035252326344646, + "grad_norm": 1.3326247930526733, + "learning_rate": 0.0001131008919263292, + "loss": 0.2021, + "step": 33760 + }, + { + "epoch": 1.3039113479284914, + "grad_norm": 1.148047685623169, + "learning_rate": 0.00011307515090672742, + "loss": 0.3987, + "step": 33770 + }, + { + "epoch": 1.3042974632225182, + "grad_norm": 0.19721268117427826, + "learning_rate": 0.00011304940988712562, + "loss": 0.2642, + "step": 33780 + }, + { + "epoch": 1.3046835785165452, + "grad_norm": 1.4060617685317993, + "learning_rate": 0.00011302366886752385, + "loss": 0.2736, + "step": 33790 + }, + { + "epoch": 1.305069693810572, + "grad_norm": 1.0736548900604248, + "learning_rate": 0.00011299792784792207, + "loss": 0.2448, + "step": 33800 + }, + { + "epoch": 1.3054558091045987, + "grad_norm": 4.352476119995117, + "learning_rate": 0.00011297218682832027, + "loss": 0.383, + "step": 33810 + }, + { + "epoch": 1.3058419243986255, + "grad_norm": 0.2249228060245514, + "learning_rate": 0.0001129464458087185, + "loss": 0.14, + "step": 33820 + }, + { + "epoch": 1.3062280396926522, + "grad_norm": 0.4820781946182251, + "learning_rate": 0.0001129207047891167, + "loss": 0.248, + "step": 33830 + }, + { + "epoch": 1.306614154986679, + "grad_norm": 2.2983391284942627, + "learning_rate": 0.00011289496376951491, + "loss": 0.2608, + "step": 33840 + }, + { + "epoch": 1.3070002702807058, + "grad_norm": 1.3315671682357788, + "learning_rate": 0.00011286922274991314, + "loss": 0.1727, + "step": 33850 + }, + { + "epoch": 1.3073863855747327, + "grad_norm": 2.060299873352051, + "learning_rate": 0.00011284348173031134, + "loss": 0.3097, + "step": 33860 + }, + { + "epoch": 1.3077725008687593, + "grad_norm": 2.096285581588745, + "learning_rate": 0.00011281774071070957, + "loss": 0.2305, + "step": 33870 + }, + { + "epoch": 1.3081586161627863, + "grad_norm": 0.4997636675834656, + "learning_rate": 0.00011279199969110777, + "loss": 0.1993, + "step": 33880 + }, + { + "epoch": 1.308544731456813, + "grad_norm": 0.58636474609375, + "learning_rate": 0.00011276625867150599, + "loss": 0.1873, + "step": 33890 + }, + { + "epoch": 1.3089308467508398, + "grad_norm": 0.9128592610359192, + "learning_rate": 0.00011274051765190419, + "loss": 0.1885, + "step": 33900 + }, + { + "epoch": 1.3093169620448666, + "grad_norm": 2.228043794631958, + "learning_rate": 0.0001127147766323024, + "loss": 0.3649, + "step": 33910 + }, + { + "epoch": 1.3097030773388934, + "grad_norm": 1.069002389907837, + "learning_rate": 0.00011268903561270063, + "loss": 0.5454, + "step": 33920 + }, + { + "epoch": 1.3100891926329201, + "grad_norm": 0.6207597851753235, + "learning_rate": 0.00011266329459309883, + "loss": 0.2329, + "step": 33930 + }, + { + "epoch": 1.3104753079269469, + "grad_norm": 1.262247920036316, + "learning_rate": 0.00011263755357349706, + "loss": 0.3437, + "step": 33940 + }, + { + "epoch": 1.3108614232209739, + "grad_norm": 1.7429994344711304, + "learning_rate": 0.00011261181255389526, + "loss": 0.228, + "step": 33950 + }, + { + "epoch": 1.3112475385150006, + "grad_norm": 0.646900475025177, + "learning_rate": 0.00011258607153429349, + "loss": 0.3739, + "step": 33960 + }, + { + "epoch": 1.3116336538090274, + "grad_norm": 1.8228782415390015, + "learning_rate": 0.0001125603305146917, + "loss": 0.2325, + "step": 33970 + }, + { + "epoch": 1.3120197691030542, + "grad_norm": 3.539228916168213, + "learning_rate": 0.0001125345894950899, + "loss": 0.194, + "step": 33980 + }, + { + "epoch": 1.312405884397081, + "grad_norm": 1.2801135778427124, + "learning_rate": 0.00011250884847548813, + "loss": 0.3069, + "step": 33990 + }, + { + "epoch": 1.3127919996911077, + "grad_norm": 3.6265695095062256, + "learning_rate": 0.00011248310745588633, + "loss": 0.3113, + "step": 34000 + }, + { + "epoch": 1.3131781149851345, + "grad_norm": 0.07370063662528992, + "learning_rate": 0.00011245736643628455, + "loss": 0.1449, + "step": 34010 + }, + { + "epoch": 1.3135642302791615, + "grad_norm": 1.0295637845993042, + "learning_rate": 0.00011243162541668275, + "loss": 0.22, + "step": 34020 + }, + { + "epoch": 1.313950345573188, + "grad_norm": 0.8803662061691284, + "learning_rate": 0.00011240588439708098, + "loss": 0.1368, + "step": 34030 + }, + { + "epoch": 1.314336460867215, + "grad_norm": 1.6597707271575928, + "learning_rate": 0.00011238014337747919, + "loss": 0.3038, + "step": 34040 + }, + { + "epoch": 1.3147225761612418, + "grad_norm": 2.115492343902588, + "learning_rate": 0.00011235440235787739, + "loss": 0.1754, + "step": 34050 + }, + { + "epoch": 1.3151086914552685, + "grad_norm": 0.8143919706344604, + "learning_rate": 0.00011232866133827562, + "loss": 0.3764, + "step": 34060 + }, + { + "epoch": 1.3154948067492953, + "grad_norm": 0.14369767904281616, + "learning_rate": 0.00011230292031867382, + "loss": 0.1742, + "step": 34070 + }, + { + "epoch": 1.315880922043322, + "grad_norm": 1.0129845142364502, + "learning_rate": 0.00011227717929907205, + "loss": 0.1458, + "step": 34080 + }, + { + "epoch": 1.316267037337349, + "grad_norm": 2.7300291061401367, + "learning_rate": 0.00011225143827947025, + "loss": 0.3939, + "step": 34090 + }, + { + "epoch": 1.3166531526313756, + "grad_norm": 0.20205609500408173, + "learning_rate": 0.00011222569725986847, + "loss": 0.29, + "step": 34100 + }, + { + "epoch": 1.3170392679254026, + "grad_norm": 1.8928464651107788, + "learning_rate": 0.00011219995624026669, + "loss": 0.1742, + "step": 34110 + }, + { + "epoch": 1.3174253832194294, + "grad_norm": 0.2639687955379486, + "learning_rate": 0.00011217421522066488, + "loss": 0.1745, + "step": 34120 + }, + { + "epoch": 1.3178114985134561, + "grad_norm": 0.5906389355659485, + "learning_rate": 0.00011214847420106311, + "loss": 0.2134, + "step": 34130 + }, + { + "epoch": 1.318197613807483, + "grad_norm": 0.9190629720687866, + "learning_rate": 0.00011212273318146131, + "loss": 0.2547, + "step": 34140 + }, + { + "epoch": 1.3185837291015097, + "grad_norm": 0.5749151110649109, + "learning_rate": 0.00011209699216185954, + "loss": 0.1688, + "step": 34150 + }, + { + "epoch": 1.3189698443955364, + "grad_norm": 0.82295823097229, + "learning_rate": 0.00011207125114225777, + "loss": 0.2884, + "step": 34160 + }, + { + "epoch": 1.3193559596895632, + "grad_norm": 0.07816460728645325, + "learning_rate": 0.00011204551012265597, + "loss": 0.2418, + "step": 34170 + }, + { + "epoch": 1.3197420749835902, + "grad_norm": 0.6417407393455505, + "learning_rate": 0.00011201976910305418, + "loss": 0.2557, + "step": 34180 + }, + { + "epoch": 1.320128190277617, + "grad_norm": 6.093267440795898, + "learning_rate": 0.00011199402808345238, + "loss": 0.3088, + "step": 34190 + }, + { + "epoch": 1.3205143055716437, + "grad_norm": 1.8861887454986572, + "learning_rate": 0.0001119682870638506, + "loss": 0.2204, + "step": 34200 + }, + { + "epoch": 1.3209004208656705, + "grad_norm": 2.3272714614868164, + "learning_rate": 0.0001119425460442488, + "loss": 0.2236, + "step": 34210 + }, + { + "epoch": 1.3212865361596973, + "grad_norm": 0.9608810544013977, + "learning_rate": 0.00011191680502464703, + "loss": 0.1897, + "step": 34220 + }, + { + "epoch": 1.321672651453724, + "grad_norm": 1.2157350778579712, + "learning_rate": 0.00011189106400504526, + "loss": 0.1526, + "step": 34230 + }, + { + "epoch": 1.3220587667477508, + "grad_norm": 1.6684671640396118, + "learning_rate": 0.00011186532298544346, + "loss": 0.3394, + "step": 34240 + }, + { + "epoch": 1.3224448820417778, + "grad_norm": 2.0432374477386475, + "learning_rate": 0.00011183958196584167, + "loss": 0.2183, + "step": 34250 + }, + { + "epoch": 1.3228309973358043, + "grad_norm": 0.9436892867088318, + "learning_rate": 0.00011181384094623988, + "loss": 0.2947, + "step": 34260 + }, + { + "epoch": 1.3232171126298313, + "grad_norm": 0.23260092735290527, + "learning_rate": 0.0001117880999266381, + "loss": 0.114, + "step": 34270 + }, + { + "epoch": 1.323603227923858, + "grad_norm": 1.2291594743728638, + "learning_rate": 0.0001117623589070363, + "loss": 0.3145, + "step": 34280 + }, + { + "epoch": 1.3239893432178849, + "grad_norm": 0.41411107778549194, + "learning_rate": 0.00011173661788743452, + "loss": 0.2937, + "step": 34290 + }, + { + "epoch": 1.3243754585119116, + "grad_norm": 2.354405164718628, + "learning_rate": 0.00011171087686783275, + "loss": 0.3933, + "step": 34300 + }, + { + "epoch": 1.3247615738059384, + "grad_norm": 2.6997978687286377, + "learning_rate": 0.00011168513584823095, + "loss": 0.1494, + "step": 34310 + }, + { + "epoch": 1.3251476890999654, + "grad_norm": 2.8430919647216797, + "learning_rate": 0.00011165939482862916, + "loss": 0.2869, + "step": 34320 + }, + { + "epoch": 1.325533804393992, + "grad_norm": 1.1737356185913086, + "learning_rate": 0.00011163365380902738, + "loss": 0.2792, + "step": 34330 + }, + { + "epoch": 1.325919919688019, + "grad_norm": 4.123973846435547, + "learning_rate": 0.00011160791278942559, + "loss": 0.5211, + "step": 34340 + }, + { + "epoch": 1.3263060349820457, + "grad_norm": 0.8862038850784302, + "learning_rate": 0.00011158217176982382, + "loss": 0.2976, + "step": 34350 + }, + { + "epoch": 1.3266921502760725, + "grad_norm": 1.8690590858459473, + "learning_rate": 0.00011155643075022202, + "loss": 0.2485, + "step": 34360 + }, + { + "epoch": 1.3270782655700992, + "grad_norm": 2.885589599609375, + "learning_rate": 0.00011153068973062024, + "loss": 0.3084, + "step": 34370 + }, + { + "epoch": 1.327464380864126, + "grad_norm": 0.9898788928985596, + "learning_rate": 0.00011150494871101844, + "loss": 0.261, + "step": 34380 + }, + { + "epoch": 1.3278504961581528, + "grad_norm": 0.6879653930664062, + "learning_rate": 0.00011147920769141667, + "loss": 0.2082, + "step": 34390 + }, + { + "epoch": 1.3282366114521795, + "grad_norm": 1.2619003057479858, + "learning_rate": 0.00011145346667181487, + "loss": 0.2402, + "step": 34400 + }, + { + "epoch": 1.3286227267462065, + "grad_norm": 1.1212007999420166, + "learning_rate": 0.00011142772565221308, + "loss": 0.3525, + "step": 34410 + }, + { + "epoch": 1.3290088420402333, + "grad_norm": 1.8431956768035889, + "learning_rate": 0.00011140198463261131, + "loss": 0.2212, + "step": 34420 + }, + { + "epoch": 1.32939495733426, + "grad_norm": 0.6185423731803894, + "learning_rate": 0.00011137624361300951, + "loss": 0.2455, + "step": 34430 + }, + { + "epoch": 1.3297810726282868, + "grad_norm": 2.3791301250457764, + "learning_rate": 0.00011135050259340774, + "loss": 0.1763, + "step": 34440 + }, + { + "epoch": 1.3301671879223136, + "grad_norm": 0.4928603768348694, + "learning_rate": 0.00011132476157380594, + "loss": 0.2381, + "step": 34450 + }, + { + "epoch": 1.3305533032163404, + "grad_norm": 1.5636029243469238, + "learning_rate": 0.00011129902055420416, + "loss": 0.1368, + "step": 34460 + }, + { + "epoch": 1.3309394185103671, + "grad_norm": 0.9425283074378967, + "learning_rate": 0.00011127327953460238, + "loss": 0.2825, + "step": 34470 + }, + { + "epoch": 1.3313255338043941, + "grad_norm": 1.2257115840911865, + "learning_rate": 0.00011124753851500058, + "loss": 0.2547, + "step": 34480 + }, + { + "epoch": 1.3317116490984207, + "grad_norm": 0.9416170716285706, + "learning_rate": 0.0001112217974953988, + "loss": 0.2766, + "step": 34490 + }, + { + "epoch": 1.3320977643924476, + "grad_norm": 0.5123847126960754, + "learning_rate": 0.000111196056475797, + "loss": 0.4733, + "step": 34500 + }, + { + "epoch": 1.3324838796864744, + "grad_norm": 1.5581384897232056, + "learning_rate": 0.00011117031545619523, + "loss": 0.1597, + "step": 34510 + }, + { + "epoch": 1.3328699949805012, + "grad_norm": 2.377333879470825, + "learning_rate": 0.00011114457443659343, + "loss": 0.209, + "step": 34520 + }, + { + "epoch": 1.333256110274528, + "grad_norm": 1.7840913534164429, + "learning_rate": 0.00011111883341699166, + "loss": 0.1759, + "step": 34530 + }, + { + "epoch": 1.3336422255685547, + "grad_norm": 1.1825993061065674, + "learning_rate": 0.00011109309239738987, + "loss": 0.2464, + "step": 34540 + }, + { + "epoch": 1.3340283408625815, + "grad_norm": 1.8859659433364868, + "learning_rate": 0.00011106735137778807, + "loss": 0.3539, + "step": 34550 + }, + { + "epoch": 1.3344144561566083, + "grad_norm": 1.9698175191879272, + "learning_rate": 0.0001110416103581863, + "loss": 0.3301, + "step": 34560 + }, + { + "epoch": 1.3348005714506352, + "grad_norm": 0.7649385333061218, + "learning_rate": 0.0001110158693385845, + "loss": 0.232, + "step": 34570 + }, + { + "epoch": 1.335186686744662, + "grad_norm": 0.56386399269104, + "learning_rate": 0.00011099012831898272, + "loss": 0.3425, + "step": 34580 + }, + { + "epoch": 1.3355728020386888, + "grad_norm": 2.956003189086914, + "learning_rate": 0.00011096438729938092, + "loss": 0.1518, + "step": 34590 + }, + { + "epoch": 1.3359589173327155, + "grad_norm": 2.612029552459717, + "learning_rate": 0.00011093864627977915, + "loss": 0.2765, + "step": 34600 + }, + { + "epoch": 1.3363450326267423, + "grad_norm": 0.9674397706985474, + "learning_rate": 0.00011091290526017736, + "loss": 0.303, + "step": 34610 + }, + { + "epoch": 1.336731147920769, + "grad_norm": 0.9578921794891357, + "learning_rate": 0.00011088716424057556, + "loss": 0.1405, + "step": 34620 + }, + { + "epoch": 1.3371172632147958, + "grad_norm": 2.168065071105957, + "learning_rate": 0.00011086142322097379, + "loss": 0.1914, + "step": 34630 + }, + { + "epoch": 1.3375033785088228, + "grad_norm": 1.3166526556015015, + "learning_rate": 0.00011083568220137199, + "loss": 0.4134, + "step": 34640 + }, + { + "epoch": 1.3378894938028496, + "grad_norm": 0.9082283973693848, + "learning_rate": 0.00011080994118177022, + "loss": 0.2693, + "step": 34650 + }, + { + "epoch": 1.3382756090968764, + "grad_norm": 2.203007698059082, + "learning_rate": 0.00011078420016216844, + "loss": 0.1847, + "step": 34660 + }, + { + "epoch": 1.3386617243909031, + "grad_norm": 0.8101674914360046, + "learning_rate": 0.00011075845914256664, + "loss": 0.3111, + "step": 34670 + }, + { + "epoch": 1.33904783968493, + "grad_norm": 1.9545695781707764, + "learning_rate": 0.00011073271812296486, + "loss": 0.3761, + "step": 34680 + }, + { + "epoch": 1.3394339549789567, + "grad_norm": 1.547581672668457, + "learning_rate": 0.00011070697710336306, + "loss": 0.2374, + "step": 34690 + }, + { + "epoch": 1.3398200702729834, + "grad_norm": 3.3519034385681152, + "learning_rate": 0.00011068123608376128, + "loss": 0.1957, + "step": 34700 + }, + { + "epoch": 1.3402061855670104, + "grad_norm": 1.5508599281311035, + "learning_rate": 0.00011065549506415948, + "loss": 0.4171, + "step": 34710 + }, + { + "epoch": 1.340592300861037, + "grad_norm": 1.8547546863555908, + "learning_rate": 0.00011062975404455771, + "loss": 0.1872, + "step": 34720 + }, + { + "epoch": 1.340978416155064, + "grad_norm": 1.4600756168365479, + "learning_rate": 0.00011060401302495594, + "loss": 0.3515, + "step": 34730 + }, + { + "epoch": 1.3413645314490907, + "grad_norm": 0.05774044618010521, + "learning_rate": 0.00011057827200535414, + "loss": 0.1604, + "step": 34740 + }, + { + "epoch": 1.3417506467431175, + "grad_norm": 2.8793342113494873, + "learning_rate": 0.00011055253098575235, + "loss": 0.3095, + "step": 34750 + }, + { + "epoch": 1.3421367620371443, + "grad_norm": 2.241042375564575, + "learning_rate": 0.00011052678996615055, + "loss": 0.2511, + "step": 34760 + }, + { + "epoch": 1.342522877331171, + "grad_norm": 1.9320632219314575, + "learning_rate": 0.00011050104894654878, + "loss": 0.4493, + "step": 34770 + }, + { + "epoch": 1.3429089926251978, + "grad_norm": 1.6483882665634155, + "learning_rate": 0.000110475307926947, + "loss": 0.217, + "step": 34780 + }, + { + "epoch": 1.3432951079192246, + "grad_norm": 0.9635765552520752, + "learning_rate": 0.0001104495669073452, + "loss": 0.5458, + "step": 34790 + }, + { + "epoch": 1.3436812232132516, + "grad_norm": 1.2436567544937134, + "learning_rate": 0.00011042382588774343, + "loss": 0.2857, + "step": 34800 + }, + { + "epoch": 1.3440673385072783, + "grad_norm": 2.8082425594329834, + "learning_rate": 0.00011039808486814163, + "loss": 0.3439, + "step": 34810 + }, + { + "epoch": 1.344453453801305, + "grad_norm": 1.0430901050567627, + "learning_rate": 0.00011037234384853984, + "loss": 0.1404, + "step": 34820 + }, + { + "epoch": 1.3448395690953319, + "grad_norm": 1.7387149333953857, + "learning_rate": 0.00011034660282893806, + "loss": 0.395, + "step": 34830 + }, + { + "epoch": 1.3452256843893586, + "grad_norm": 1.2713748216629028, + "learning_rate": 0.00011032086180933627, + "loss": 0.29, + "step": 34840 + }, + { + "epoch": 1.3456117996833854, + "grad_norm": 0.26068204641342163, + "learning_rate": 0.0001102951207897345, + "loss": 0.1814, + "step": 34850 + }, + { + "epoch": 1.3459979149774122, + "grad_norm": 2.4163243770599365, + "learning_rate": 0.0001102693797701327, + "loss": 0.2445, + "step": 34860 + }, + { + "epoch": 1.3463840302714392, + "grad_norm": 2.2439687252044678, + "learning_rate": 0.00011024363875053092, + "loss": 0.3338, + "step": 34870 + }, + { + "epoch": 1.346770145565466, + "grad_norm": 0.2822403609752655, + "learning_rate": 0.00011021789773092912, + "loss": 0.1648, + "step": 34880 + }, + { + "epoch": 1.3471562608594927, + "grad_norm": 0.07319017499685287, + "learning_rate": 0.00011019215671132734, + "loss": 0.107, + "step": 34890 + }, + { + "epoch": 1.3475423761535195, + "grad_norm": 0.9809044003486633, + "learning_rate": 0.00011016641569172555, + "loss": 0.256, + "step": 34900 + }, + { + "epoch": 1.3479284914475462, + "grad_norm": 0.5016226768493652, + "learning_rate": 0.00011014067467212376, + "loss": 0.3229, + "step": 34910 + }, + { + "epoch": 1.348314606741573, + "grad_norm": 1.3026005029678345, + "learning_rate": 0.00011011493365252199, + "loss": 0.2145, + "step": 34920 + }, + { + "epoch": 1.3487007220355998, + "grad_norm": 1.0752215385437012, + "learning_rate": 0.00011008919263292019, + "loss": 0.2355, + "step": 34930 + }, + { + "epoch": 1.3490868373296268, + "grad_norm": 2.2703003883361816, + "learning_rate": 0.00011006345161331842, + "loss": 0.2079, + "step": 34940 + }, + { + "epoch": 1.3494729526236533, + "grad_norm": 1.1323810815811157, + "learning_rate": 0.00011003771059371662, + "loss": 0.1015, + "step": 34950 + }, + { + "epoch": 1.3498590679176803, + "grad_norm": 0.10813555121421814, + "learning_rate": 0.00011001196957411484, + "loss": 0.4214, + "step": 34960 + }, + { + "epoch": 1.350245183211707, + "grad_norm": 0.07815568149089813, + "learning_rate": 0.00010998622855451306, + "loss": 0.1211, + "step": 34970 + }, + { + "epoch": 1.3506312985057338, + "grad_norm": 0.6748234629631042, + "learning_rate": 0.00010996048753491126, + "loss": 0.3508, + "step": 34980 + }, + { + "epoch": 1.3510174137997606, + "grad_norm": 1.8556997776031494, + "learning_rate": 0.00010993474651530948, + "loss": 0.2268, + "step": 34990 + }, + { + "epoch": 1.3514035290937874, + "grad_norm": 0.8696061372756958, + "learning_rate": 0.00010990900549570768, + "loss": 0.4321, + "step": 35000 + }, + { + "epoch": 1.3517896443878141, + "grad_norm": 0.42442765831947327, + "learning_rate": 0.00010988326447610591, + "loss": 0.1944, + "step": 35010 + }, + { + "epoch": 1.352175759681841, + "grad_norm": 1.0474554300308228, + "learning_rate": 0.00010985752345650411, + "loss": 0.1342, + "step": 35020 + }, + { + "epoch": 1.3525618749758679, + "grad_norm": 0.607037365436554, + "learning_rate": 0.00010983178243690234, + "loss": 0.2965, + "step": 35030 + }, + { + "epoch": 1.3529479902698947, + "grad_norm": 1.8160990476608276, + "learning_rate": 0.00010980604141730055, + "loss": 0.3192, + "step": 35040 + }, + { + "epoch": 1.3533341055639214, + "grad_norm": 2.0026509761810303, + "learning_rate": 0.00010978030039769875, + "loss": 0.3054, + "step": 35050 + }, + { + "epoch": 1.3537202208579482, + "grad_norm": 0.9203600883483887, + "learning_rate": 0.00010975455937809698, + "loss": 0.253, + "step": 35060 + }, + { + "epoch": 1.354106336151975, + "grad_norm": 0.33198195695877075, + "learning_rate": 0.00010972881835849518, + "loss": 0.3885, + "step": 35070 + }, + { + "epoch": 1.3544924514460017, + "grad_norm": 0.3201223611831665, + "learning_rate": 0.0001097030773388934, + "loss": 0.3029, + "step": 35080 + }, + { + "epoch": 1.3548785667400285, + "grad_norm": 1.2589943408966064, + "learning_rate": 0.0001096773363192916, + "loss": 0.4243, + "step": 35090 + }, + { + "epoch": 1.3552646820340555, + "grad_norm": 1.5106219053268433, + "learning_rate": 0.00010965159529968983, + "loss": 0.2585, + "step": 35100 + }, + { + "epoch": 1.3556507973280822, + "grad_norm": 1.429799199104309, + "learning_rate": 0.00010962585428008804, + "loss": 0.1961, + "step": 35110 + }, + { + "epoch": 1.356036912622109, + "grad_norm": 2.1211297512054443, + "learning_rate": 0.00010960011326048624, + "loss": 0.4057, + "step": 35120 + }, + { + "epoch": 1.3564230279161358, + "grad_norm": 2.5154731273651123, + "learning_rate": 0.00010957437224088447, + "loss": 0.3787, + "step": 35130 + }, + { + "epoch": 1.3568091432101625, + "grad_norm": 0.4914834201335907, + "learning_rate": 0.00010954863122128267, + "loss": 0.234, + "step": 35140 + }, + { + "epoch": 1.3571952585041893, + "grad_norm": 0.26685893535614014, + "learning_rate": 0.0001095228902016809, + "loss": 0.2841, + "step": 35150 + }, + { + "epoch": 1.357581373798216, + "grad_norm": 0.15462155640125275, + "learning_rate": 0.00010949714918207912, + "loss": 0.2269, + "step": 35160 + }, + { + "epoch": 1.357967489092243, + "grad_norm": 1.3887063264846802, + "learning_rate": 0.00010947140816247732, + "loss": 0.3455, + "step": 35170 + }, + { + "epoch": 1.3583536043862696, + "grad_norm": 0.786374032497406, + "learning_rate": 0.00010944566714287554, + "loss": 0.2897, + "step": 35180 + }, + { + "epoch": 1.3587397196802966, + "grad_norm": 1.100475549697876, + "learning_rate": 0.00010941992612327374, + "loss": 0.2892, + "step": 35190 + }, + { + "epoch": 1.3591258349743234, + "grad_norm": 0.7676102519035339, + "learning_rate": 0.00010939418510367196, + "loss": 0.1942, + "step": 35200 + }, + { + "epoch": 1.3595119502683501, + "grad_norm": 0.33462053537368774, + "learning_rate": 0.00010936844408407016, + "loss": 0.2872, + "step": 35210 + }, + { + "epoch": 1.359898065562377, + "grad_norm": 0.9294387698173523, + "learning_rate": 0.00010934270306446839, + "loss": 0.2617, + "step": 35220 + }, + { + "epoch": 1.3602841808564037, + "grad_norm": 0.3169979453086853, + "learning_rate": 0.00010931696204486662, + "loss": 0.2942, + "step": 35230 + }, + { + "epoch": 1.3606702961504304, + "grad_norm": 2.1339616775512695, + "learning_rate": 0.00010929122102526482, + "loss": 0.4448, + "step": 35240 + }, + { + "epoch": 1.3610564114444572, + "grad_norm": 0.9430062770843506, + "learning_rate": 0.00010926548000566303, + "loss": 0.2051, + "step": 35250 + }, + { + "epoch": 1.3614425267384842, + "grad_norm": 3.1187360286712646, + "learning_rate": 0.00010923973898606123, + "loss": 0.2274, + "step": 35260 + }, + { + "epoch": 1.361828642032511, + "grad_norm": 1.4727579355239868, + "learning_rate": 0.00010921399796645946, + "loss": 0.3757, + "step": 35270 + }, + { + "epoch": 1.3622147573265377, + "grad_norm": 2.157560348510742, + "learning_rate": 0.00010918825694685768, + "loss": 0.3096, + "step": 35280 + }, + { + "epoch": 1.3626008726205645, + "grad_norm": 0.33457377552986145, + "learning_rate": 0.00010916251592725588, + "loss": 0.1489, + "step": 35290 + }, + { + "epoch": 1.3629869879145913, + "grad_norm": 0.9005904197692871, + "learning_rate": 0.00010913677490765411, + "loss": 0.1826, + "step": 35300 + }, + { + "epoch": 1.363373103208618, + "grad_norm": 2.1222829818725586, + "learning_rate": 0.00010911103388805231, + "loss": 0.1965, + "step": 35310 + }, + { + "epoch": 1.3637592185026448, + "grad_norm": 1.3881357908248901, + "learning_rate": 0.00010908529286845052, + "loss": 0.1791, + "step": 35320 + }, + { + "epoch": 1.3641453337966718, + "grad_norm": 1.7574503421783447, + "learning_rate": 0.00010905955184884872, + "loss": 0.3316, + "step": 35330 + }, + { + "epoch": 1.3645314490906983, + "grad_norm": 0.1967727392911911, + "learning_rate": 0.00010903381082924695, + "loss": 0.2331, + "step": 35340 + }, + { + "epoch": 1.3649175643847253, + "grad_norm": 0.8974360823631287, + "learning_rate": 0.00010900806980964518, + "loss": 0.2589, + "step": 35350 + }, + { + "epoch": 1.365303679678752, + "grad_norm": 2.0996744632720947, + "learning_rate": 0.00010898232879004338, + "loss": 0.3663, + "step": 35360 + }, + { + "epoch": 1.3656897949727789, + "grad_norm": 0.5678316354751587, + "learning_rate": 0.0001089565877704416, + "loss": 0.1729, + "step": 35370 + }, + { + "epoch": 1.3660759102668056, + "grad_norm": 2.3381874561309814, + "learning_rate": 0.0001089308467508398, + "loss": 0.1615, + "step": 35380 + }, + { + "epoch": 1.3664620255608324, + "grad_norm": 1.0276836156845093, + "learning_rate": 0.00010890510573123802, + "loss": 0.3359, + "step": 35390 + }, + { + "epoch": 1.3668481408548594, + "grad_norm": 2.4374940395355225, + "learning_rate": 0.00010887936471163622, + "loss": 0.2435, + "step": 35400 + }, + { + "epoch": 1.367234256148886, + "grad_norm": 0.45221665501594543, + "learning_rate": 0.00010885362369203444, + "loss": 0.2555, + "step": 35410 + }, + { + "epoch": 1.367620371442913, + "grad_norm": 2.608090400695801, + "learning_rate": 0.00010882788267243267, + "loss": 0.2465, + "step": 35420 + }, + { + "epoch": 1.3680064867369397, + "grad_norm": 1.4186642169952393, + "learning_rate": 0.00010880214165283087, + "loss": 0.1674, + "step": 35430 + }, + { + "epoch": 1.3683926020309665, + "grad_norm": 0.659479022026062, + "learning_rate": 0.0001087764006332291, + "loss": 0.2926, + "step": 35440 + }, + { + "epoch": 1.3687787173249932, + "grad_norm": 0.9219567179679871, + "learning_rate": 0.0001087506596136273, + "loss": 0.2001, + "step": 35450 + }, + { + "epoch": 1.36916483261902, + "grad_norm": 0.8070804476737976, + "learning_rate": 0.00010872491859402551, + "loss": 0.2178, + "step": 35460 + }, + { + "epoch": 1.3695509479130468, + "grad_norm": 2.9981069564819336, + "learning_rate": 0.00010869917757442374, + "loss": 0.3079, + "step": 35470 + }, + { + "epoch": 1.3699370632070735, + "grad_norm": 0.7891242504119873, + "learning_rate": 0.00010867343655482194, + "loss": 0.2765, + "step": 35480 + }, + { + "epoch": 1.3703231785011005, + "grad_norm": 1.448637843132019, + "learning_rate": 0.00010864769553522016, + "loss": 0.3521, + "step": 35490 + }, + { + "epoch": 1.3707092937951273, + "grad_norm": 0.07628043740987778, + "learning_rate": 0.00010862195451561836, + "loss": 0.2083, + "step": 35500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.161800511488e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3e515339fee736e2bc683eda3c364abb3f9fef56 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5cd4ddf0ae551cf34064bc7ce8cfe5e8c5878d139f8bfea61ff6b6141b4a2895 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..d2afeabe1ef55d796655b679b39c99b2e0257a50 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14db796de6976a3a4b92f5111673e38a999f191d8c45d13648e0db4e38dd1daf +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..75e980090d57b58ea4a894f9b17308c0bbe8d820 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..96bf8498b40ed496d03523ace8267d7864a1c9da Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..422ca72162846b4b7cae1a2de10ffb4bdca2d095 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/trainer_state.json @@ -0,0 +1,25234 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.390015058496467, + "eval_steps": 500, + "global_step": 36000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + }, + { + "epoch": 1.274566585582455, + "grad_norm": 0.5019121766090393, + "learning_rate": 0.0001150314683964632, + "loss": 0.1857, + "step": 33010 + }, + { + "epoch": 1.2749527008764816, + "grad_norm": 0.258952260017395, + "learning_rate": 0.0001150057273768614, + "loss": 0.11, + "step": 33020 + }, + { + "epoch": 1.2753388161705086, + "grad_norm": 0.8540348410606384, + "learning_rate": 0.00011497998635725962, + "loss": 0.1852, + "step": 33030 + }, + { + "epoch": 1.2757249314645354, + "grad_norm": 0.08689398318529129, + "learning_rate": 0.00011495424533765783, + "loss": 0.2225, + "step": 33040 + }, + { + "epoch": 1.2761110467585621, + "grad_norm": 0.42253822088241577, + "learning_rate": 0.00011492850431805605, + "loss": 0.0751, + "step": 33050 + }, + { + "epoch": 1.276497162052589, + "grad_norm": 1.2964017391204834, + "learning_rate": 0.00011490276329845426, + "loss": 0.2384, + "step": 33060 + }, + { + "epoch": 1.2768832773466157, + "grad_norm": 0.5337836146354675, + "learning_rate": 0.00011487702227885246, + "loss": 0.1415, + "step": 33070 + }, + { + "epoch": 1.2772693926406424, + "grad_norm": 2.7771682739257812, + "learning_rate": 0.00011485128125925069, + "loss": 0.378, + "step": 33080 + }, + { + "epoch": 1.2776555079346692, + "grad_norm": 1.5107232332229614, + "learning_rate": 0.00011482554023964889, + "loss": 0.2482, + "step": 33090 + }, + { + "epoch": 1.2780416232286962, + "grad_norm": 0.6885499358177185, + "learning_rate": 0.00011479979922004711, + "loss": 0.2122, + "step": 33100 + }, + { + "epoch": 1.278427738522723, + "grad_norm": 0.9016557335853577, + "learning_rate": 0.00011477405820044533, + "loss": 0.2841, + "step": 33110 + }, + { + "epoch": 1.2788138538167497, + "grad_norm": 1.9532525539398193, + "learning_rate": 0.00011474831718084354, + "loss": 0.2281, + "step": 33120 + }, + { + "epoch": 1.2791999691107765, + "grad_norm": 2.1078782081604004, + "learning_rate": 0.00011472257616124175, + "loss": 0.2836, + "step": 33130 + }, + { + "epoch": 1.2795860844048033, + "grad_norm": 0.19830390810966492, + "learning_rate": 0.00011469683514163995, + "loss": 0.388, + "step": 33140 + }, + { + "epoch": 1.27997219969883, + "grad_norm": 0.17538850009441376, + "learning_rate": 0.00011467109412203818, + "loss": 0.3274, + "step": 33150 + }, + { + "epoch": 1.2803583149928568, + "grad_norm": 0.7402139902114868, + "learning_rate": 0.00011464535310243641, + "loss": 0.1979, + "step": 33160 + }, + { + "epoch": 1.2807444302868838, + "grad_norm": 0.2097146362066269, + "learning_rate": 0.00011461961208283461, + "loss": 0.2464, + "step": 33170 + }, + { + "epoch": 1.2811305455809103, + "grad_norm": 1.2441083192825317, + "learning_rate": 0.00011459387106323283, + "loss": 0.266, + "step": 33180 + }, + { + "epoch": 1.2815166608749373, + "grad_norm": 2.518852710723877, + "learning_rate": 0.00011456813004363103, + "loss": 0.253, + "step": 33190 + }, + { + "epoch": 1.281902776168964, + "grad_norm": 0.8078998327255249, + "learning_rate": 0.00011454238902402925, + "loss": 0.2361, + "step": 33200 + }, + { + "epoch": 1.2822888914629909, + "grad_norm": 1.2297371625900269, + "learning_rate": 0.00011451664800442745, + "loss": 0.1974, + "step": 33210 + }, + { + "epoch": 1.2826750067570176, + "grad_norm": 0.4303855895996094, + "learning_rate": 0.00011449090698482567, + "loss": 0.3563, + "step": 33220 + }, + { + "epoch": 1.2830611220510444, + "grad_norm": 1.3215210437774658, + "learning_rate": 0.0001144651659652239, + "loss": 0.2818, + "step": 33230 + }, + { + "epoch": 1.2834472373450712, + "grad_norm": 1.546265959739685, + "learning_rate": 0.0001144394249456221, + "loss": 0.5778, + "step": 33240 + }, + { + "epoch": 1.283833352639098, + "grad_norm": 0.8895953297615051, + "learning_rate": 0.00011441368392602033, + "loss": 0.2485, + "step": 33250 + }, + { + "epoch": 1.284219467933125, + "grad_norm": 0.7534870505332947, + "learning_rate": 0.00011438794290641853, + "loss": 0.2825, + "step": 33260 + }, + { + "epoch": 1.2846055832271517, + "grad_norm": 0.052820973098278046, + "learning_rate": 0.00011436220188681674, + "loss": 0.2191, + "step": 33270 + }, + { + "epoch": 1.2849916985211784, + "grad_norm": 0.9264475107192993, + "learning_rate": 0.00011433646086721494, + "loss": 0.181, + "step": 33280 + }, + { + "epoch": 1.2853778138152052, + "grad_norm": 0.2128441333770752, + "learning_rate": 0.00011431071984761317, + "loss": 0.1819, + "step": 33290 + }, + { + "epoch": 1.285763929109232, + "grad_norm": 0.5400950312614441, + "learning_rate": 0.0001142849788280114, + "loss": 0.4316, + "step": 33300 + }, + { + "epoch": 1.2861500444032588, + "grad_norm": 0.8033271431922913, + "learning_rate": 0.00011425923780840959, + "loss": 0.2146, + "step": 33310 + }, + { + "epoch": 1.2865361596972855, + "grad_norm": 2.012575149536133, + "learning_rate": 0.00011423349678880782, + "loss": 0.4335, + "step": 33320 + }, + { + "epoch": 1.2869222749913125, + "grad_norm": 0.7352376580238342, + "learning_rate": 0.00011420775576920602, + "loss": 0.2124, + "step": 33330 + }, + { + "epoch": 1.2873083902853393, + "grad_norm": 0.769036591053009, + "learning_rate": 0.00011418201474960423, + "loss": 0.3602, + "step": 33340 + }, + { + "epoch": 1.287694505579366, + "grad_norm": 0.250592976808548, + "learning_rate": 0.00011415627373000246, + "loss": 0.1692, + "step": 33350 + }, + { + "epoch": 1.2880806208733928, + "grad_norm": 2.43820858001709, + "learning_rate": 0.00011413053271040066, + "loss": 0.2777, + "step": 33360 + }, + { + "epoch": 1.2884667361674196, + "grad_norm": 1.3179954290390015, + "learning_rate": 0.00011410479169079889, + "loss": 0.1794, + "step": 33370 + }, + { + "epoch": 1.2888528514614463, + "grad_norm": 1.0040466785430908, + "learning_rate": 0.00011407905067119709, + "loss": 0.3037, + "step": 33380 + }, + { + "epoch": 1.2892389667554731, + "grad_norm": 5.296288013458252, + "learning_rate": 0.00011405330965159531, + "loss": 0.2904, + "step": 33390 + }, + { + "epoch": 1.2896250820495, + "grad_norm": 0.4267273247241974, + "learning_rate": 0.00011402756863199351, + "loss": 0.2263, + "step": 33400 + }, + { + "epoch": 1.2900111973435266, + "grad_norm": 0.8817713260650635, + "learning_rate": 0.00011400182761239173, + "loss": 0.2715, + "step": 33410 + }, + { + "epoch": 1.2903973126375536, + "grad_norm": 2.7891275882720947, + "learning_rate": 0.00011397608659278995, + "loss": 0.2781, + "step": 33420 + }, + { + "epoch": 1.2907834279315804, + "grad_norm": 0.3013952374458313, + "learning_rate": 0.00011395034557318815, + "loss": 0.2563, + "step": 33430 + }, + { + "epoch": 1.2911695432256072, + "grad_norm": 1.766413927078247, + "learning_rate": 0.00011392460455358638, + "loss": 0.1854, + "step": 33440 + }, + { + "epoch": 1.291555658519634, + "grad_norm": 0.25331103801727295, + "learning_rate": 0.00011389886353398458, + "loss": 0.1206, + "step": 33450 + }, + { + "epoch": 1.2919417738136607, + "grad_norm": 0.036400288343429565, + "learning_rate": 0.0001138731225143828, + "loss": 0.4707, + "step": 33460 + }, + { + "epoch": 1.2923278891076875, + "grad_norm": 1.5620888471603394, + "learning_rate": 0.00011384738149478102, + "loss": 0.3312, + "step": 33470 + }, + { + "epoch": 1.2927140044017142, + "grad_norm": 0.6670392155647278, + "learning_rate": 0.00011382164047517922, + "loss": 0.2341, + "step": 33480 + }, + { + "epoch": 1.2931001196957412, + "grad_norm": 2.3108737468719482, + "learning_rate": 0.00011379589945557745, + "loss": 0.3843, + "step": 33490 + }, + { + "epoch": 1.293486234989768, + "grad_norm": 0.8025147318840027, + "learning_rate": 0.00011377015843597565, + "loss": 0.1982, + "step": 33500 + }, + { + "epoch": 1.2938723502837948, + "grad_norm": 1.7835719585418701, + "learning_rate": 0.00011374441741637387, + "loss": 0.3285, + "step": 33510 + }, + { + "epoch": 1.2942584655778215, + "grad_norm": 2.041508913040161, + "learning_rate": 0.00011371867639677207, + "loss": 0.2044, + "step": 33520 + }, + { + "epoch": 1.2946445808718483, + "grad_norm": 1.103378415107727, + "learning_rate": 0.0001136929353771703, + "loss": 0.1682, + "step": 33530 + }, + { + "epoch": 1.295030696165875, + "grad_norm": 0.057376351207494736, + "learning_rate": 0.00011366719435756851, + "loss": 0.1642, + "step": 33540 + }, + { + "epoch": 1.2954168114599018, + "grad_norm": 0.6539410948753357, + "learning_rate": 0.00011364145333796671, + "loss": 0.1549, + "step": 33550 + }, + { + "epoch": 1.2958029267539288, + "grad_norm": 1.250543236732483, + "learning_rate": 0.00011361571231836494, + "loss": 0.3764, + "step": 33560 + }, + { + "epoch": 1.2961890420479556, + "grad_norm": 0.23697887361049652, + "learning_rate": 0.00011358997129876314, + "loss": 0.3999, + "step": 33570 + }, + { + "epoch": 1.2965751573419824, + "grad_norm": 0.9318505525588989, + "learning_rate": 0.00011356423027916137, + "loss": 0.4156, + "step": 33580 + }, + { + "epoch": 1.2969612726360091, + "grad_norm": 1.3910777568817139, + "learning_rate": 0.00011353848925955957, + "loss": 0.3455, + "step": 33590 + }, + { + "epoch": 1.297347387930036, + "grad_norm": 1.6764451265335083, + "learning_rate": 0.00011351274823995779, + "loss": 0.1884, + "step": 33600 + }, + { + "epoch": 1.2977335032240627, + "grad_norm": 0.9300051927566528, + "learning_rate": 0.000113487007220356, + "loss": 0.119, + "step": 33610 + }, + { + "epoch": 1.2981196185180894, + "grad_norm": 2.447462558746338, + "learning_rate": 0.00011346126620075422, + "loss": 0.4403, + "step": 33620 + }, + { + "epoch": 1.2985057338121164, + "grad_norm": 1.216407060623169, + "learning_rate": 0.00011343552518115243, + "loss": 0.2415, + "step": 33630 + }, + { + "epoch": 1.298891849106143, + "grad_norm": 2.968648910522461, + "learning_rate": 0.00011340978416155063, + "loss": 0.2899, + "step": 33640 + }, + { + "epoch": 1.29927796440017, + "grad_norm": 0.6649970412254333, + "learning_rate": 0.00011338404314194886, + "loss": 0.3809, + "step": 33650 + }, + { + "epoch": 1.2996640796941967, + "grad_norm": 1.7277917861938477, + "learning_rate": 0.00011335830212234709, + "loss": 0.3308, + "step": 33660 + }, + { + "epoch": 1.3000501949882235, + "grad_norm": 1.3269709348678589, + "learning_rate": 0.00011333256110274529, + "loss": 0.3682, + "step": 33670 + }, + { + "epoch": 1.3004363102822503, + "grad_norm": 0.20609407126903534, + "learning_rate": 0.0001133068200831435, + "loss": 0.1379, + "step": 33680 + }, + { + "epoch": 1.300822425576277, + "grad_norm": 0.6592215299606323, + "learning_rate": 0.00011328107906354171, + "loss": 0.2746, + "step": 33690 + }, + { + "epoch": 1.3012085408703038, + "grad_norm": 1.903635859489441, + "learning_rate": 0.00011325533804393993, + "loss": 0.4729, + "step": 33700 + }, + { + "epoch": 1.3015946561643306, + "grad_norm": 0.8432504534721375, + "learning_rate": 0.00011322959702433813, + "loss": 0.2835, + "step": 33710 + }, + { + "epoch": 1.3019807714583576, + "grad_norm": 0.9862542152404785, + "learning_rate": 0.00011320385600473635, + "loss": 0.1823, + "step": 33720 + }, + { + "epoch": 1.3023668867523843, + "grad_norm": 3.845738649368286, + "learning_rate": 0.00011317811498513458, + "loss": 0.2542, + "step": 33730 + }, + { + "epoch": 1.302753002046411, + "grad_norm": 0.6317747235298157, + "learning_rate": 0.00011315237396553278, + "loss": 0.22, + "step": 33740 + }, + { + "epoch": 1.3031391173404379, + "grad_norm": 2.5221354961395264, + "learning_rate": 0.000113126632945931, + "loss": 0.2253, + "step": 33750 + }, + { + "epoch": 1.3035252326344646, + "grad_norm": 1.3326247930526733, + "learning_rate": 0.0001131008919263292, + "loss": 0.2021, + "step": 33760 + }, + { + "epoch": 1.3039113479284914, + "grad_norm": 1.148047685623169, + "learning_rate": 0.00011307515090672742, + "loss": 0.3987, + "step": 33770 + }, + { + "epoch": 1.3042974632225182, + "grad_norm": 0.19721268117427826, + "learning_rate": 0.00011304940988712562, + "loss": 0.2642, + "step": 33780 + }, + { + "epoch": 1.3046835785165452, + "grad_norm": 1.4060617685317993, + "learning_rate": 0.00011302366886752385, + "loss": 0.2736, + "step": 33790 + }, + { + "epoch": 1.305069693810572, + "grad_norm": 1.0736548900604248, + "learning_rate": 0.00011299792784792207, + "loss": 0.2448, + "step": 33800 + }, + { + "epoch": 1.3054558091045987, + "grad_norm": 4.352476119995117, + "learning_rate": 0.00011297218682832027, + "loss": 0.383, + "step": 33810 + }, + { + "epoch": 1.3058419243986255, + "grad_norm": 0.2249228060245514, + "learning_rate": 0.0001129464458087185, + "loss": 0.14, + "step": 33820 + }, + { + "epoch": 1.3062280396926522, + "grad_norm": 0.4820781946182251, + "learning_rate": 0.0001129207047891167, + "loss": 0.248, + "step": 33830 + }, + { + "epoch": 1.306614154986679, + "grad_norm": 2.2983391284942627, + "learning_rate": 0.00011289496376951491, + "loss": 0.2608, + "step": 33840 + }, + { + "epoch": 1.3070002702807058, + "grad_norm": 1.3315671682357788, + "learning_rate": 0.00011286922274991314, + "loss": 0.1727, + "step": 33850 + }, + { + "epoch": 1.3073863855747327, + "grad_norm": 2.060299873352051, + "learning_rate": 0.00011284348173031134, + "loss": 0.3097, + "step": 33860 + }, + { + "epoch": 1.3077725008687593, + "grad_norm": 2.096285581588745, + "learning_rate": 0.00011281774071070957, + "loss": 0.2305, + "step": 33870 + }, + { + "epoch": 1.3081586161627863, + "grad_norm": 0.4997636675834656, + "learning_rate": 0.00011279199969110777, + "loss": 0.1993, + "step": 33880 + }, + { + "epoch": 1.308544731456813, + "grad_norm": 0.58636474609375, + "learning_rate": 0.00011276625867150599, + "loss": 0.1873, + "step": 33890 + }, + { + "epoch": 1.3089308467508398, + "grad_norm": 0.9128592610359192, + "learning_rate": 0.00011274051765190419, + "loss": 0.1885, + "step": 33900 + }, + { + "epoch": 1.3093169620448666, + "grad_norm": 2.228043794631958, + "learning_rate": 0.0001127147766323024, + "loss": 0.3649, + "step": 33910 + }, + { + "epoch": 1.3097030773388934, + "grad_norm": 1.069002389907837, + "learning_rate": 0.00011268903561270063, + "loss": 0.5454, + "step": 33920 + }, + { + "epoch": 1.3100891926329201, + "grad_norm": 0.6207597851753235, + "learning_rate": 0.00011266329459309883, + "loss": 0.2329, + "step": 33930 + }, + { + "epoch": 1.3104753079269469, + "grad_norm": 1.262247920036316, + "learning_rate": 0.00011263755357349706, + "loss": 0.3437, + "step": 33940 + }, + { + "epoch": 1.3108614232209739, + "grad_norm": 1.7429994344711304, + "learning_rate": 0.00011261181255389526, + "loss": 0.228, + "step": 33950 + }, + { + "epoch": 1.3112475385150006, + "grad_norm": 0.646900475025177, + "learning_rate": 0.00011258607153429349, + "loss": 0.3739, + "step": 33960 + }, + { + "epoch": 1.3116336538090274, + "grad_norm": 1.8228782415390015, + "learning_rate": 0.0001125603305146917, + "loss": 0.2325, + "step": 33970 + }, + { + "epoch": 1.3120197691030542, + "grad_norm": 3.539228916168213, + "learning_rate": 0.0001125345894950899, + "loss": 0.194, + "step": 33980 + }, + { + "epoch": 1.312405884397081, + "grad_norm": 1.2801135778427124, + "learning_rate": 0.00011250884847548813, + "loss": 0.3069, + "step": 33990 + }, + { + "epoch": 1.3127919996911077, + "grad_norm": 3.6265695095062256, + "learning_rate": 0.00011248310745588633, + "loss": 0.3113, + "step": 34000 + }, + { + "epoch": 1.3131781149851345, + "grad_norm": 0.07370063662528992, + "learning_rate": 0.00011245736643628455, + "loss": 0.1449, + "step": 34010 + }, + { + "epoch": 1.3135642302791615, + "grad_norm": 1.0295637845993042, + "learning_rate": 0.00011243162541668275, + "loss": 0.22, + "step": 34020 + }, + { + "epoch": 1.313950345573188, + "grad_norm": 0.8803662061691284, + "learning_rate": 0.00011240588439708098, + "loss": 0.1368, + "step": 34030 + }, + { + "epoch": 1.314336460867215, + "grad_norm": 1.6597707271575928, + "learning_rate": 0.00011238014337747919, + "loss": 0.3038, + "step": 34040 + }, + { + "epoch": 1.3147225761612418, + "grad_norm": 2.115492343902588, + "learning_rate": 0.00011235440235787739, + "loss": 0.1754, + "step": 34050 + }, + { + "epoch": 1.3151086914552685, + "grad_norm": 0.8143919706344604, + "learning_rate": 0.00011232866133827562, + "loss": 0.3764, + "step": 34060 + }, + { + "epoch": 1.3154948067492953, + "grad_norm": 0.14369767904281616, + "learning_rate": 0.00011230292031867382, + "loss": 0.1742, + "step": 34070 + }, + { + "epoch": 1.315880922043322, + "grad_norm": 1.0129845142364502, + "learning_rate": 0.00011227717929907205, + "loss": 0.1458, + "step": 34080 + }, + { + "epoch": 1.316267037337349, + "grad_norm": 2.7300291061401367, + "learning_rate": 0.00011225143827947025, + "loss": 0.3939, + "step": 34090 + }, + { + "epoch": 1.3166531526313756, + "grad_norm": 0.20205609500408173, + "learning_rate": 0.00011222569725986847, + "loss": 0.29, + "step": 34100 + }, + { + "epoch": 1.3170392679254026, + "grad_norm": 1.8928464651107788, + "learning_rate": 0.00011219995624026669, + "loss": 0.1742, + "step": 34110 + }, + { + "epoch": 1.3174253832194294, + "grad_norm": 0.2639687955379486, + "learning_rate": 0.00011217421522066488, + "loss": 0.1745, + "step": 34120 + }, + { + "epoch": 1.3178114985134561, + "grad_norm": 0.5906389355659485, + "learning_rate": 0.00011214847420106311, + "loss": 0.2134, + "step": 34130 + }, + { + "epoch": 1.318197613807483, + "grad_norm": 0.9190629720687866, + "learning_rate": 0.00011212273318146131, + "loss": 0.2547, + "step": 34140 + }, + { + "epoch": 1.3185837291015097, + "grad_norm": 0.5749151110649109, + "learning_rate": 0.00011209699216185954, + "loss": 0.1688, + "step": 34150 + }, + { + "epoch": 1.3189698443955364, + "grad_norm": 0.82295823097229, + "learning_rate": 0.00011207125114225777, + "loss": 0.2884, + "step": 34160 + }, + { + "epoch": 1.3193559596895632, + "grad_norm": 0.07816460728645325, + "learning_rate": 0.00011204551012265597, + "loss": 0.2418, + "step": 34170 + }, + { + "epoch": 1.3197420749835902, + "grad_norm": 0.6417407393455505, + "learning_rate": 0.00011201976910305418, + "loss": 0.2557, + "step": 34180 + }, + { + "epoch": 1.320128190277617, + "grad_norm": 6.093267440795898, + "learning_rate": 0.00011199402808345238, + "loss": 0.3088, + "step": 34190 + }, + { + "epoch": 1.3205143055716437, + "grad_norm": 1.8861887454986572, + "learning_rate": 0.0001119682870638506, + "loss": 0.2204, + "step": 34200 + }, + { + "epoch": 1.3209004208656705, + "grad_norm": 2.3272714614868164, + "learning_rate": 0.0001119425460442488, + "loss": 0.2236, + "step": 34210 + }, + { + "epoch": 1.3212865361596973, + "grad_norm": 0.9608810544013977, + "learning_rate": 0.00011191680502464703, + "loss": 0.1897, + "step": 34220 + }, + { + "epoch": 1.321672651453724, + "grad_norm": 1.2157350778579712, + "learning_rate": 0.00011189106400504526, + "loss": 0.1526, + "step": 34230 + }, + { + "epoch": 1.3220587667477508, + "grad_norm": 1.6684671640396118, + "learning_rate": 0.00011186532298544346, + "loss": 0.3394, + "step": 34240 + }, + { + "epoch": 1.3224448820417778, + "grad_norm": 2.0432374477386475, + "learning_rate": 0.00011183958196584167, + "loss": 0.2183, + "step": 34250 + }, + { + "epoch": 1.3228309973358043, + "grad_norm": 0.9436892867088318, + "learning_rate": 0.00011181384094623988, + "loss": 0.2947, + "step": 34260 + }, + { + "epoch": 1.3232171126298313, + "grad_norm": 0.23260092735290527, + "learning_rate": 0.0001117880999266381, + "loss": 0.114, + "step": 34270 + }, + { + "epoch": 1.323603227923858, + "grad_norm": 1.2291594743728638, + "learning_rate": 0.0001117623589070363, + "loss": 0.3145, + "step": 34280 + }, + { + "epoch": 1.3239893432178849, + "grad_norm": 0.41411107778549194, + "learning_rate": 0.00011173661788743452, + "loss": 0.2937, + "step": 34290 + }, + { + "epoch": 1.3243754585119116, + "grad_norm": 2.354405164718628, + "learning_rate": 0.00011171087686783275, + "loss": 0.3933, + "step": 34300 + }, + { + "epoch": 1.3247615738059384, + "grad_norm": 2.6997978687286377, + "learning_rate": 0.00011168513584823095, + "loss": 0.1494, + "step": 34310 + }, + { + "epoch": 1.3251476890999654, + "grad_norm": 2.8430919647216797, + "learning_rate": 0.00011165939482862916, + "loss": 0.2869, + "step": 34320 + }, + { + "epoch": 1.325533804393992, + "grad_norm": 1.1737356185913086, + "learning_rate": 0.00011163365380902738, + "loss": 0.2792, + "step": 34330 + }, + { + "epoch": 1.325919919688019, + "grad_norm": 4.123973846435547, + "learning_rate": 0.00011160791278942559, + "loss": 0.5211, + "step": 34340 + }, + { + "epoch": 1.3263060349820457, + "grad_norm": 0.8862038850784302, + "learning_rate": 0.00011158217176982382, + "loss": 0.2976, + "step": 34350 + }, + { + "epoch": 1.3266921502760725, + "grad_norm": 1.8690590858459473, + "learning_rate": 0.00011155643075022202, + "loss": 0.2485, + "step": 34360 + }, + { + "epoch": 1.3270782655700992, + "grad_norm": 2.885589599609375, + "learning_rate": 0.00011153068973062024, + "loss": 0.3084, + "step": 34370 + }, + { + "epoch": 1.327464380864126, + "grad_norm": 0.9898788928985596, + "learning_rate": 0.00011150494871101844, + "loss": 0.261, + "step": 34380 + }, + { + "epoch": 1.3278504961581528, + "grad_norm": 0.6879653930664062, + "learning_rate": 0.00011147920769141667, + "loss": 0.2082, + "step": 34390 + }, + { + "epoch": 1.3282366114521795, + "grad_norm": 1.2619003057479858, + "learning_rate": 0.00011145346667181487, + "loss": 0.2402, + "step": 34400 + }, + { + "epoch": 1.3286227267462065, + "grad_norm": 1.1212007999420166, + "learning_rate": 0.00011142772565221308, + "loss": 0.3525, + "step": 34410 + }, + { + "epoch": 1.3290088420402333, + "grad_norm": 1.8431956768035889, + "learning_rate": 0.00011140198463261131, + "loss": 0.2212, + "step": 34420 + }, + { + "epoch": 1.32939495733426, + "grad_norm": 0.6185423731803894, + "learning_rate": 0.00011137624361300951, + "loss": 0.2455, + "step": 34430 + }, + { + "epoch": 1.3297810726282868, + "grad_norm": 2.3791301250457764, + "learning_rate": 0.00011135050259340774, + "loss": 0.1763, + "step": 34440 + }, + { + "epoch": 1.3301671879223136, + "grad_norm": 0.4928603768348694, + "learning_rate": 0.00011132476157380594, + "loss": 0.2381, + "step": 34450 + }, + { + "epoch": 1.3305533032163404, + "grad_norm": 1.5636029243469238, + "learning_rate": 0.00011129902055420416, + "loss": 0.1368, + "step": 34460 + }, + { + "epoch": 1.3309394185103671, + "grad_norm": 0.9425283074378967, + "learning_rate": 0.00011127327953460238, + "loss": 0.2825, + "step": 34470 + }, + { + "epoch": 1.3313255338043941, + "grad_norm": 1.2257115840911865, + "learning_rate": 0.00011124753851500058, + "loss": 0.2547, + "step": 34480 + }, + { + "epoch": 1.3317116490984207, + "grad_norm": 0.9416170716285706, + "learning_rate": 0.0001112217974953988, + "loss": 0.2766, + "step": 34490 + }, + { + "epoch": 1.3320977643924476, + "grad_norm": 0.5123847126960754, + "learning_rate": 0.000111196056475797, + "loss": 0.4733, + "step": 34500 + }, + { + "epoch": 1.3324838796864744, + "grad_norm": 1.5581384897232056, + "learning_rate": 0.00011117031545619523, + "loss": 0.1597, + "step": 34510 + }, + { + "epoch": 1.3328699949805012, + "grad_norm": 2.377333879470825, + "learning_rate": 0.00011114457443659343, + "loss": 0.209, + "step": 34520 + }, + { + "epoch": 1.333256110274528, + "grad_norm": 1.7840913534164429, + "learning_rate": 0.00011111883341699166, + "loss": 0.1759, + "step": 34530 + }, + { + "epoch": 1.3336422255685547, + "grad_norm": 1.1825993061065674, + "learning_rate": 0.00011109309239738987, + "loss": 0.2464, + "step": 34540 + }, + { + "epoch": 1.3340283408625815, + "grad_norm": 1.8859659433364868, + "learning_rate": 0.00011106735137778807, + "loss": 0.3539, + "step": 34550 + }, + { + "epoch": 1.3344144561566083, + "grad_norm": 1.9698175191879272, + "learning_rate": 0.0001110416103581863, + "loss": 0.3301, + "step": 34560 + }, + { + "epoch": 1.3348005714506352, + "grad_norm": 0.7649385333061218, + "learning_rate": 0.0001110158693385845, + "loss": 0.232, + "step": 34570 + }, + { + "epoch": 1.335186686744662, + "grad_norm": 0.56386399269104, + "learning_rate": 0.00011099012831898272, + "loss": 0.3425, + "step": 34580 + }, + { + "epoch": 1.3355728020386888, + "grad_norm": 2.956003189086914, + "learning_rate": 0.00011096438729938092, + "loss": 0.1518, + "step": 34590 + }, + { + "epoch": 1.3359589173327155, + "grad_norm": 2.612029552459717, + "learning_rate": 0.00011093864627977915, + "loss": 0.2765, + "step": 34600 + }, + { + "epoch": 1.3363450326267423, + "grad_norm": 0.9674397706985474, + "learning_rate": 0.00011091290526017736, + "loss": 0.303, + "step": 34610 + }, + { + "epoch": 1.336731147920769, + "grad_norm": 0.9578921794891357, + "learning_rate": 0.00011088716424057556, + "loss": 0.1405, + "step": 34620 + }, + { + "epoch": 1.3371172632147958, + "grad_norm": 2.168065071105957, + "learning_rate": 0.00011086142322097379, + "loss": 0.1914, + "step": 34630 + }, + { + "epoch": 1.3375033785088228, + "grad_norm": 1.3166526556015015, + "learning_rate": 0.00011083568220137199, + "loss": 0.4134, + "step": 34640 + }, + { + "epoch": 1.3378894938028496, + "grad_norm": 0.9082283973693848, + "learning_rate": 0.00011080994118177022, + "loss": 0.2693, + "step": 34650 + }, + { + "epoch": 1.3382756090968764, + "grad_norm": 2.203007698059082, + "learning_rate": 0.00011078420016216844, + "loss": 0.1847, + "step": 34660 + }, + { + "epoch": 1.3386617243909031, + "grad_norm": 0.8101674914360046, + "learning_rate": 0.00011075845914256664, + "loss": 0.3111, + "step": 34670 + }, + { + "epoch": 1.33904783968493, + "grad_norm": 1.9545695781707764, + "learning_rate": 0.00011073271812296486, + "loss": 0.3761, + "step": 34680 + }, + { + "epoch": 1.3394339549789567, + "grad_norm": 1.547581672668457, + "learning_rate": 0.00011070697710336306, + "loss": 0.2374, + "step": 34690 + }, + { + "epoch": 1.3398200702729834, + "grad_norm": 3.3519034385681152, + "learning_rate": 0.00011068123608376128, + "loss": 0.1957, + "step": 34700 + }, + { + "epoch": 1.3402061855670104, + "grad_norm": 1.5508599281311035, + "learning_rate": 0.00011065549506415948, + "loss": 0.4171, + "step": 34710 + }, + { + "epoch": 1.340592300861037, + "grad_norm": 1.8547546863555908, + "learning_rate": 0.00011062975404455771, + "loss": 0.1872, + "step": 34720 + }, + { + "epoch": 1.340978416155064, + "grad_norm": 1.4600756168365479, + "learning_rate": 0.00011060401302495594, + "loss": 0.3515, + "step": 34730 + }, + { + "epoch": 1.3413645314490907, + "grad_norm": 0.05774044618010521, + "learning_rate": 0.00011057827200535414, + "loss": 0.1604, + "step": 34740 + }, + { + "epoch": 1.3417506467431175, + "grad_norm": 2.8793342113494873, + "learning_rate": 0.00011055253098575235, + "loss": 0.3095, + "step": 34750 + }, + { + "epoch": 1.3421367620371443, + "grad_norm": 2.241042375564575, + "learning_rate": 0.00011052678996615055, + "loss": 0.2511, + "step": 34760 + }, + { + "epoch": 1.342522877331171, + "grad_norm": 1.9320632219314575, + "learning_rate": 0.00011050104894654878, + "loss": 0.4493, + "step": 34770 + }, + { + "epoch": 1.3429089926251978, + "grad_norm": 1.6483882665634155, + "learning_rate": 0.000110475307926947, + "loss": 0.217, + "step": 34780 + }, + { + "epoch": 1.3432951079192246, + "grad_norm": 0.9635765552520752, + "learning_rate": 0.0001104495669073452, + "loss": 0.5458, + "step": 34790 + }, + { + "epoch": 1.3436812232132516, + "grad_norm": 1.2436567544937134, + "learning_rate": 0.00011042382588774343, + "loss": 0.2857, + "step": 34800 + }, + { + "epoch": 1.3440673385072783, + "grad_norm": 2.8082425594329834, + "learning_rate": 0.00011039808486814163, + "loss": 0.3439, + "step": 34810 + }, + { + "epoch": 1.344453453801305, + "grad_norm": 1.0430901050567627, + "learning_rate": 0.00011037234384853984, + "loss": 0.1404, + "step": 34820 + }, + { + "epoch": 1.3448395690953319, + "grad_norm": 1.7387149333953857, + "learning_rate": 0.00011034660282893806, + "loss": 0.395, + "step": 34830 + }, + { + "epoch": 1.3452256843893586, + "grad_norm": 1.2713748216629028, + "learning_rate": 0.00011032086180933627, + "loss": 0.29, + "step": 34840 + }, + { + "epoch": 1.3456117996833854, + "grad_norm": 0.26068204641342163, + "learning_rate": 0.0001102951207897345, + "loss": 0.1814, + "step": 34850 + }, + { + "epoch": 1.3459979149774122, + "grad_norm": 2.4163243770599365, + "learning_rate": 0.0001102693797701327, + "loss": 0.2445, + "step": 34860 + }, + { + "epoch": 1.3463840302714392, + "grad_norm": 2.2439687252044678, + "learning_rate": 0.00011024363875053092, + "loss": 0.3338, + "step": 34870 + }, + { + "epoch": 1.346770145565466, + "grad_norm": 0.2822403609752655, + "learning_rate": 0.00011021789773092912, + "loss": 0.1648, + "step": 34880 + }, + { + "epoch": 1.3471562608594927, + "grad_norm": 0.07319017499685287, + "learning_rate": 0.00011019215671132734, + "loss": 0.107, + "step": 34890 + }, + { + "epoch": 1.3475423761535195, + "grad_norm": 0.9809044003486633, + "learning_rate": 0.00011016641569172555, + "loss": 0.256, + "step": 34900 + }, + { + "epoch": 1.3479284914475462, + "grad_norm": 0.5016226768493652, + "learning_rate": 0.00011014067467212376, + "loss": 0.3229, + "step": 34910 + }, + { + "epoch": 1.348314606741573, + "grad_norm": 1.3026005029678345, + "learning_rate": 0.00011011493365252199, + "loss": 0.2145, + "step": 34920 + }, + { + "epoch": 1.3487007220355998, + "grad_norm": 1.0752215385437012, + "learning_rate": 0.00011008919263292019, + "loss": 0.2355, + "step": 34930 + }, + { + "epoch": 1.3490868373296268, + "grad_norm": 2.2703003883361816, + "learning_rate": 0.00011006345161331842, + "loss": 0.2079, + "step": 34940 + }, + { + "epoch": 1.3494729526236533, + "grad_norm": 1.1323810815811157, + "learning_rate": 0.00011003771059371662, + "loss": 0.1015, + "step": 34950 + }, + { + "epoch": 1.3498590679176803, + "grad_norm": 0.10813555121421814, + "learning_rate": 0.00011001196957411484, + "loss": 0.4214, + "step": 34960 + }, + { + "epoch": 1.350245183211707, + "grad_norm": 0.07815568149089813, + "learning_rate": 0.00010998622855451306, + "loss": 0.1211, + "step": 34970 + }, + { + "epoch": 1.3506312985057338, + "grad_norm": 0.6748234629631042, + "learning_rate": 0.00010996048753491126, + "loss": 0.3508, + "step": 34980 + }, + { + "epoch": 1.3510174137997606, + "grad_norm": 1.8556997776031494, + "learning_rate": 0.00010993474651530948, + "loss": 0.2268, + "step": 34990 + }, + { + "epoch": 1.3514035290937874, + "grad_norm": 0.8696061372756958, + "learning_rate": 0.00010990900549570768, + "loss": 0.4321, + "step": 35000 + }, + { + "epoch": 1.3517896443878141, + "grad_norm": 0.42442765831947327, + "learning_rate": 0.00010988326447610591, + "loss": 0.1944, + "step": 35010 + }, + { + "epoch": 1.352175759681841, + "grad_norm": 1.0474554300308228, + "learning_rate": 0.00010985752345650411, + "loss": 0.1342, + "step": 35020 + }, + { + "epoch": 1.3525618749758679, + "grad_norm": 0.607037365436554, + "learning_rate": 0.00010983178243690234, + "loss": 0.2965, + "step": 35030 + }, + { + "epoch": 1.3529479902698947, + "grad_norm": 1.8160990476608276, + "learning_rate": 0.00010980604141730055, + "loss": 0.3192, + "step": 35040 + }, + { + "epoch": 1.3533341055639214, + "grad_norm": 2.0026509761810303, + "learning_rate": 0.00010978030039769875, + "loss": 0.3054, + "step": 35050 + }, + { + "epoch": 1.3537202208579482, + "grad_norm": 0.9203600883483887, + "learning_rate": 0.00010975455937809698, + "loss": 0.253, + "step": 35060 + }, + { + "epoch": 1.354106336151975, + "grad_norm": 0.33198195695877075, + "learning_rate": 0.00010972881835849518, + "loss": 0.3885, + "step": 35070 + }, + { + "epoch": 1.3544924514460017, + "grad_norm": 0.3201223611831665, + "learning_rate": 0.0001097030773388934, + "loss": 0.3029, + "step": 35080 + }, + { + "epoch": 1.3548785667400285, + "grad_norm": 1.2589943408966064, + "learning_rate": 0.0001096773363192916, + "loss": 0.4243, + "step": 35090 + }, + { + "epoch": 1.3552646820340555, + "grad_norm": 1.5106219053268433, + "learning_rate": 0.00010965159529968983, + "loss": 0.2585, + "step": 35100 + }, + { + "epoch": 1.3556507973280822, + "grad_norm": 1.429799199104309, + "learning_rate": 0.00010962585428008804, + "loss": 0.1961, + "step": 35110 + }, + { + "epoch": 1.356036912622109, + "grad_norm": 2.1211297512054443, + "learning_rate": 0.00010960011326048624, + "loss": 0.4057, + "step": 35120 + }, + { + "epoch": 1.3564230279161358, + "grad_norm": 2.5154731273651123, + "learning_rate": 0.00010957437224088447, + "loss": 0.3787, + "step": 35130 + }, + { + "epoch": 1.3568091432101625, + "grad_norm": 0.4914834201335907, + "learning_rate": 0.00010954863122128267, + "loss": 0.234, + "step": 35140 + }, + { + "epoch": 1.3571952585041893, + "grad_norm": 0.26685893535614014, + "learning_rate": 0.0001095228902016809, + "loss": 0.2841, + "step": 35150 + }, + { + "epoch": 1.357581373798216, + "grad_norm": 0.15462155640125275, + "learning_rate": 0.00010949714918207912, + "loss": 0.2269, + "step": 35160 + }, + { + "epoch": 1.357967489092243, + "grad_norm": 1.3887063264846802, + "learning_rate": 0.00010947140816247732, + "loss": 0.3455, + "step": 35170 + }, + { + "epoch": 1.3583536043862696, + "grad_norm": 0.786374032497406, + "learning_rate": 0.00010944566714287554, + "loss": 0.2897, + "step": 35180 + }, + { + "epoch": 1.3587397196802966, + "grad_norm": 1.100475549697876, + "learning_rate": 0.00010941992612327374, + "loss": 0.2892, + "step": 35190 + }, + { + "epoch": 1.3591258349743234, + "grad_norm": 0.7676102519035339, + "learning_rate": 0.00010939418510367196, + "loss": 0.1942, + "step": 35200 + }, + { + "epoch": 1.3595119502683501, + "grad_norm": 0.33462053537368774, + "learning_rate": 0.00010936844408407016, + "loss": 0.2872, + "step": 35210 + }, + { + "epoch": 1.359898065562377, + "grad_norm": 0.9294387698173523, + "learning_rate": 0.00010934270306446839, + "loss": 0.2617, + "step": 35220 + }, + { + "epoch": 1.3602841808564037, + "grad_norm": 0.3169979453086853, + "learning_rate": 0.00010931696204486662, + "loss": 0.2942, + "step": 35230 + }, + { + "epoch": 1.3606702961504304, + "grad_norm": 2.1339616775512695, + "learning_rate": 0.00010929122102526482, + "loss": 0.4448, + "step": 35240 + }, + { + "epoch": 1.3610564114444572, + "grad_norm": 0.9430062770843506, + "learning_rate": 0.00010926548000566303, + "loss": 0.2051, + "step": 35250 + }, + { + "epoch": 1.3614425267384842, + "grad_norm": 3.1187360286712646, + "learning_rate": 0.00010923973898606123, + "loss": 0.2274, + "step": 35260 + }, + { + "epoch": 1.361828642032511, + "grad_norm": 1.4727579355239868, + "learning_rate": 0.00010921399796645946, + "loss": 0.3757, + "step": 35270 + }, + { + "epoch": 1.3622147573265377, + "grad_norm": 2.157560348510742, + "learning_rate": 0.00010918825694685768, + "loss": 0.3096, + "step": 35280 + }, + { + "epoch": 1.3626008726205645, + "grad_norm": 0.33457377552986145, + "learning_rate": 0.00010916251592725588, + "loss": 0.1489, + "step": 35290 + }, + { + "epoch": 1.3629869879145913, + "grad_norm": 0.9005904197692871, + "learning_rate": 0.00010913677490765411, + "loss": 0.1826, + "step": 35300 + }, + { + "epoch": 1.363373103208618, + "grad_norm": 2.1222829818725586, + "learning_rate": 0.00010911103388805231, + "loss": 0.1965, + "step": 35310 + }, + { + "epoch": 1.3637592185026448, + "grad_norm": 1.3881357908248901, + "learning_rate": 0.00010908529286845052, + "loss": 0.1791, + "step": 35320 + }, + { + "epoch": 1.3641453337966718, + "grad_norm": 1.7574503421783447, + "learning_rate": 0.00010905955184884872, + "loss": 0.3316, + "step": 35330 + }, + { + "epoch": 1.3645314490906983, + "grad_norm": 0.1967727392911911, + "learning_rate": 0.00010903381082924695, + "loss": 0.2331, + "step": 35340 + }, + { + "epoch": 1.3649175643847253, + "grad_norm": 0.8974360823631287, + "learning_rate": 0.00010900806980964518, + "loss": 0.2589, + "step": 35350 + }, + { + "epoch": 1.365303679678752, + "grad_norm": 2.0996744632720947, + "learning_rate": 0.00010898232879004338, + "loss": 0.3663, + "step": 35360 + }, + { + "epoch": 1.3656897949727789, + "grad_norm": 0.5678316354751587, + "learning_rate": 0.0001089565877704416, + "loss": 0.1729, + "step": 35370 + }, + { + "epoch": 1.3660759102668056, + "grad_norm": 2.3381874561309814, + "learning_rate": 0.0001089308467508398, + "loss": 0.1615, + "step": 35380 + }, + { + "epoch": 1.3664620255608324, + "grad_norm": 1.0276836156845093, + "learning_rate": 0.00010890510573123802, + "loss": 0.3359, + "step": 35390 + }, + { + "epoch": 1.3668481408548594, + "grad_norm": 2.4374940395355225, + "learning_rate": 0.00010887936471163622, + "loss": 0.2435, + "step": 35400 + }, + { + "epoch": 1.367234256148886, + "grad_norm": 0.45221665501594543, + "learning_rate": 0.00010885362369203444, + "loss": 0.2555, + "step": 35410 + }, + { + "epoch": 1.367620371442913, + "grad_norm": 2.608090400695801, + "learning_rate": 0.00010882788267243267, + "loss": 0.2465, + "step": 35420 + }, + { + "epoch": 1.3680064867369397, + "grad_norm": 1.4186642169952393, + "learning_rate": 0.00010880214165283087, + "loss": 0.1674, + "step": 35430 + }, + { + "epoch": 1.3683926020309665, + "grad_norm": 0.659479022026062, + "learning_rate": 0.0001087764006332291, + "loss": 0.2926, + "step": 35440 + }, + { + "epoch": 1.3687787173249932, + "grad_norm": 0.9219567179679871, + "learning_rate": 0.0001087506596136273, + "loss": 0.2001, + "step": 35450 + }, + { + "epoch": 1.36916483261902, + "grad_norm": 0.8070804476737976, + "learning_rate": 0.00010872491859402551, + "loss": 0.2178, + "step": 35460 + }, + { + "epoch": 1.3695509479130468, + "grad_norm": 2.9981069564819336, + "learning_rate": 0.00010869917757442374, + "loss": 0.3079, + "step": 35470 + }, + { + "epoch": 1.3699370632070735, + "grad_norm": 0.7891242504119873, + "learning_rate": 0.00010867343655482194, + "loss": 0.2765, + "step": 35480 + }, + { + "epoch": 1.3703231785011005, + "grad_norm": 1.448637843132019, + "learning_rate": 0.00010864769553522016, + "loss": 0.3521, + "step": 35490 + }, + { + "epoch": 1.3707092937951273, + "grad_norm": 0.07628043740987778, + "learning_rate": 0.00010862195451561836, + "loss": 0.2083, + "step": 35500 + }, + { + "epoch": 1.371095409089154, + "grad_norm": 0.7549735307693481, + "learning_rate": 0.00010859621349601659, + "loss": 0.2536, + "step": 35510 + }, + { + "epoch": 1.3714815243831808, + "grad_norm": 1.3548041582107544, + "learning_rate": 0.00010857047247641479, + "loss": 0.251, + "step": 35520 + }, + { + "epoch": 1.3718676396772076, + "grad_norm": 0.530010998249054, + "learning_rate": 0.000108544731456813, + "loss": 0.1917, + "step": 35530 + }, + { + "epoch": 1.3722537549712344, + "grad_norm": 0.4148992896080017, + "learning_rate": 0.00010851899043721123, + "loss": 0.335, + "step": 35540 + }, + { + "epoch": 1.3726398702652611, + "grad_norm": 1.5118776559829712, + "learning_rate": 0.00010849324941760943, + "loss": 0.2159, + "step": 35550 + }, + { + "epoch": 1.3730259855592881, + "grad_norm": 1.036889910697937, + "learning_rate": 0.00010846750839800766, + "loss": 0.2975, + "step": 35560 + }, + { + "epoch": 1.3734121008533147, + "grad_norm": 1.724263072013855, + "learning_rate": 0.00010844176737840586, + "loss": 0.1476, + "step": 35570 + }, + { + "epoch": 1.3737982161473417, + "grad_norm": 1.599007487297058, + "learning_rate": 0.00010841602635880408, + "loss": 0.2539, + "step": 35580 + }, + { + "epoch": 1.3741843314413684, + "grad_norm": 2.9119279384613037, + "learning_rate": 0.00010839028533920228, + "loss": 0.2688, + "step": 35590 + }, + { + "epoch": 1.3745704467353952, + "grad_norm": 1.8647874593734741, + "learning_rate": 0.00010836454431960051, + "loss": 0.4158, + "step": 35600 + }, + { + "epoch": 1.374956562029422, + "grad_norm": 3.925290822982788, + "learning_rate": 0.00010833880329999872, + "loss": 0.3333, + "step": 35610 + }, + { + "epoch": 1.3753426773234487, + "grad_norm": 0.7124634385108948, + "learning_rate": 0.00010831306228039692, + "loss": 0.1069, + "step": 35620 + }, + { + "epoch": 1.3757287926174757, + "grad_norm": 1.303579330444336, + "learning_rate": 0.00010828732126079515, + "loss": 0.2898, + "step": 35630 + }, + { + "epoch": 1.3761149079115023, + "grad_norm": 3.921804189682007, + "learning_rate": 0.00010826158024119335, + "loss": 0.4212, + "step": 35640 + }, + { + "epoch": 1.3765010232055293, + "grad_norm": 1.3194564580917358, + "learning_rate": 0.00010823583922159158, + "loss": 0.2771, + "step": 35650 + }, + { + "epoch": 1.376887138499556, + "grad_norm": 1.4237637519836426, + "learning_rate": 0.00010821009820198979, + "loss": 0.2463, + "step": 35660 + }, + { + "epoch": 1.3772732537935828, + "grad_norm": 1.8165888786315918, + "learning_rate": 0.000108184357182388, + "loss": 0.291, + "step": 35670 + }, + { + "epoch": 1.3776593690876096, + "grad_norm": 1.1056426763534546, + "learning_rate": 0.00010815861616278622, + "loss": 0.2525, + "step": 35680 + }, + { + "epoch": 1.3780454843816363, + "grad_norm": 1.483189582824707, + "learning_rate": 0.00010813287514318442, + "loss": 0.1569, + "step": 35690 + }, + { + "epoch": 1.378431599675663, + "grad_norm": 1.0666841268539429, + "learning_rate": 0.00010810713412358264, + "loss": 0.235, + "step": 35700 + }, + { + "epoch": 1.3788177149696899, + "grad_norm": 1.0299845933914185, + "learning_rate": 0.00010808139310398084, + "loss": 0.3892, + "step": 35710 + }, + { + "epoch": 1.3792038302637168, + "grad_norm": 2.3474409580230713, + "learning_rate": 0.00010805565208437907, + "loss": 0.3417, + "step": 35720 + }, + { + "epoch": 1.3795899455577436, + "grad_norm": 1.7456315755844116, + "learning_rate": 0.0001080299110647773, + "loss": 0.2538, + "step": 35730 + }, + { + "epoch": 1.3799760608517704, + "grad_norm": 2.866103410720825, + "learning_rate": 0.0001080041700451755, + "loss": 0.1619, + "step": 35740 + }, + { + "epoch": 1.3803621761457971, + "grad_norm": 0.29136407375335693, + "learning_rate": 0.00010797842902557371, + "loss": 0.2692, + "step": 35750 + }, + { + "epoch": 1.380748291439824, + "grad_norm": 0.8046161532402039, + "learning_rate": 0.00010795268800597191, + "loss": 0.1575, + "step": 35760 + }, + { + "epoch": 1.3811344067338507, + "grad_norm": 0.6451787352561951, + "learning_rate": 0.00010792694698637014, + "loss": 0.4914, + "step": 35770 + }, + { + "epoch": 1.3815205220278775, + "grad_norm": 0.7289161086082458, + "learning_rate": 0.00010790120596676836, + "loss": 0.1895, + "step": 35780 + }, + { + "epoch": 1.3819066373219044, + "grad_norm": 0.8300430178642273, + "learning_rate": 0.00010787546494716656, + "loss": 0.3663, + "step": 35790 + }, + { + "epoch": 1.382292752615931, + "grad_norm": 0.17713364958763123, + "learning_rate": 0.00010784972392756479, + "loss": 0.3189, + "step": 35800 + }, + { + "epoch": 1.382678867909958, + "grad_norm": 0.903222918510437, + "learning_rate": 0.00010782398290796299, + "loss": 0.1577, + "step": 35810 + }, + { + "epoch": 1.3830649832039847, + "grad_norm": 0.08617932349443436, + "learning_rate": 0.0001077982418883612, + "loss": 0.2872, + "step": 35820 + }, + { + "epoch": 1.3834510984980115, + "grad_norm": 1.9590895175933838, + "learning_rate": 0.0001077725008687594, + "loss": 0.2907, + "step": 35830 + }, + { + "epoch": 1.3838372137920383, + "grad_norm": 1.2515161037445068, + "learning_rate": 0.00010774675984915763, + "loss": 0.177, + "step": 35840 + }, + { + "epoch": 1.384223329086065, + "grad_norm": 1.6171292066574097, + "learning_rate": 0.00010772101882955586, + "loss": 0.2321, + "step": 35850 + }, + { + "epoch": 1.3846094443800918, + "grad_norm": 0.13681405782699585, + "learning_rate": 0.00010769527780995406, + "loss": 0.247, + "step": 35860 + }, + { + "epoch": 1.3849955596741186, + "grad_norm": 1.1949968338012695, + "learning_rate": 0.00010766953679035228, + "loss": 0.2978, + "step": 35870 + }, + { + "epoch": 1.3853816749681456, + "grad_norm": 0.17001692950725555, + "learning_rate": 0.00010764379577075048, + "loss": 0.3241, + "step": 35880 + }, + { + "epoch": 1.3857677902621723, + "grad_norm": 0.8227952122688293, + "learning_rate": 0.0001076180547511487, + "loss": 0.3499, + "step": 35890 + }, + { + "epoch": 1.386153905556199, + "grad_norm": 1.4185482263565063, + "learning_rate": 0.0001075923137315469, + "loss": 0.3109, + "step": 35900 + }, + { + "epoch": 1.3865400208502259, + "grad_norm": 0.9533351063728333, + "learning_rate": 0.00010756657271194512, + "loss": 0.24, + "step": 35910 + }, + { + "epoch": 1.3869261361442526, + "grad_norm": 1.077789306640625, + "learning_rate": 0.00010754083169234335, + "loss": 0.2662, + "step": 35920 + }, + { + "epoch": 1.3873122514382794, + "grad_norm": 1.3528363704681396, + "learning_rate": 0.00010751509067274155, + "loss": 0.1623, + "step": 35930 + }, + { + "epoch": 1.3876983667323062, + "grad_norm": 0.25122806429862976, + "learning_rate": 0.00010748934965313978, + "loss": 0.1639, + "step": 35940 + }, + { + "epoch": 1.3880844820263332, + "grad_norm": 0.9446159601211548, + "learning_rate": 0.00010746360863353798, + "loss": 0.2035, + "step": 35950 + }, + { + "epoch": 1.38847059732036, + "grad_norm": 1.2258719205856323, + "learning_rate": 0.00010743786761393619, + "loss": 0.4247, + "step": 35960 + }, + { + "epoch": 1.3888567126143867, + "grad_norm": 3.117729663848877, + "learning_rate": 0.00010741212659433442, + "loss": 0.2569, + "step": 35970 + }, + { + "epoch": 1.3892428279084135, + "grad_norm": 0.6333123445510864, + "learning_rate": 0.00010738638557473262, + "loss": 0.1843, + "step": 35980 + }, + { + "epoch": 1.3896289432024402, + "grad_norm": 1.49360191822052, + "learning_rate": 0.00010736064455513084, + "loss": 0.2318, + "step": 35990 + }, + { + "epoch": 1.390015058496467, + "grad_norm": 3.9082753658294678, + "learning_rate": 0.00010733490353552904, + "loss": 0.2037, + "step": 36000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.192248406016e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3b277b22072e5d1ee1fd103337c02ee77cd2832e --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2899e0ea6fd7da7e45a6aaf89286cb5b78ad2a1031c6bd672a1eb51b240ab914 +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..68c95ea2cd14c3ff432ac6d5c1fd869c5434427d --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78b0ce06b0cc2df087a09381efa62f9dcd2fa05b1274db42ecf6f101fa9a472d +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..3c15b9b23572c809d337f77e427306bd25650bc2 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..1939b4070101d87bd40b17ce05762ceb525dec4c Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..f90590860018b42d819aec23e906c7305758b5f6 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/trainer_state.json @@ -0,0 +1,25584 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.409320823197807, + "eval_steps": 500, + "global_step": 36500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + }, + { + "epoch": 1.274566585582455, + "grad_norm": 0.5019121766090393, + "learning_rate": 0.0001150314683964632, + "loss": 0.1857, + "step": 33010 + }, + { + "epoch": 1.2749527008764816, + "grad_norm": 0.258952260017395, + "learning_rate": 0.0001150057273768614, + "loss": 0.11, + "step": 33020 + }, + { + "epoch": 1.2753388161705086, + "grad_norm": 0.8540348410606384, + "learning_rate": 0.00011497998635725962, + "loss": 0.1852, + "step": 33030 + }, + { + "epoch": 1.2757249314645354, + "grad_norm": 0.08689398318529129, + "learning_rate": 0.00011495424533765783, + "loss": 0.2225, + "step": 33040 + }, + { + "epoch": 1.2761110467585621, + "grad_norm": 0.42253822088241577, + "learning_rate": 0.00011492850431805605, + "loss": 0.0751, + "step": 33050 + }, + { + "epoch": 1.276497162052589, + "grad_norm": 1.2964017391204834, + "learning_rate": 0.00011490276329845426, + "loss": 0.2384, + "step": 33060 + }, + { + "epoch": 1.2768832773466157, + "grad_norm": 0.5337836146354675, + "learning_rate": 0.00011487702227885246, + "loss": 0.1415, + "step": 33070 + }, + { + "epoch": 1.2772693926406424, + "grad_norm": 2.7771682739257812, + "learning_rate": 0.00011485128125925069, + "loss": 0.378, + "step": 33080 + }, + { + "epoch": 1.2776555079346692, + "grad_norm": 1.5107232332229614, + "learning_rate": 0.00011482554023964889, + "loss": 0.2482, + "step": 33090 + }, + { + "epoch": 1.2780416232286962, + "grad_norm": 0.6885499358177185, + "learning_rate": 0.00011479979922004711, + "loss": 0.2122, + "step": 33100 + }, + { + "epoch": 1.278427738522723, + "grad_norm": 0.9016557335853577, + "learning_rate": 0.00011477405820044533, + "loss": 0.2841, + "step": 33110 + }, + { + "epoch": 1.2788138538167497, + "grad_norm": 1.9532525539398193, + "learning_rate": 0.00011474831718084354, + "loss": 0.2281, + "step": 33120 + }, + { + "epoch": 1.2791999691107765, + "grad_norm": 2.1078782081604004, + "learning_rate": 0.00011472257616124175, + "loss": 0.2836, + "step": 33130 + }, + { + "epoch": 1.2795860844048033, + "grad_norm": 0.19830390810966492, + "learning_rate": 0.00011469683514163995, + "loss": 0.388, + "step": 33140 + }, + { + "epoch": 1.27997219969883, + "grad_norm": 0.17538850009441376, + "learning_rate": 0.00011467109412203818, + "loss": 0.3274, + "step": 33150 + }, + { + "epoch": 1.2803583149928568, + "grad_norm": 0.7402139902114868, + "learning_rate": 0.00011464535310243641, + "loss": 0.1979, + "step": 33160 + }, + { + "epoch": 1.2807444302868838, + "grad_norm": 0.2097146362066269, + "learning_rate": 0.00011461961208283461, + "loss": 0.2464, + "step": 33170 + }, + { + "epoch": 1.2811305455809103, + "grad_norm": 1.2441083192825317, + "learning_rate": 0.00011459387106323283, + "loss": 0.266, + "step": 33180 + }, + { + "epoch": 1.2815166608749373, + "grad_norm": 2.518852710723877, + "learning_rate": 0.00011456813004363103, + "loss": 0.253, + "step": 33190 + }, + { + "epoch": 1.281902776168964, + "grad_norm": 0.8078998327255249, + "learning_rate": 0.00011454238902402925, + "loss": 0.2361, + "step": 33200 + }, + { + "epoch": 1.2822888914629909, + "grad_norm": 1.2297371625900269, + "learning_rate": 0.00011451664800442745, + "loss": 0.1974, + "step": 33210 + }, + { + "epoch": 1.2826750067570176, + "grad_norm": 0.4303855895996094, + "learning_rate": 0.00011449090698482567, + "loss": 0.3563, + "step": 33220 + }, + { + "epoch": 1.2830611220510444, + "grad_norm": 1.3215210437774658, + "learning_rate": 0.0001144651659652239, + "loss": 0.2818, + "step": 33230 + }, + { + "epoch": 1.2834472373450712, + "grad_norm": 1.546265959739685, + "learning_rate": 0.0001144394249456221, + "loss": 0.5778, + "step": 33240 + }, + { + "epoch": 1.283833352639098, + "grad_norm": 0.8895953297615051, + "learning_rate": 0.00011441368392602033, + "loss": 0.2485, + "step": 33250 + }, + { + "epoch": 1.284219467933125, + "grad_norm": 0.7534870505332947, + "learning_rate": 0.00011438794290641853, + "loss": 0.2825, + "step": 33260 + }, + { + "epoch": 1.2846055832271517, + "grad_norm": 0.052820973098278046, + "learning_rate": 0.00011436220188681674, + "loss": 0.2191, + "step": 33270 + }, + { + "epoch": 1.2849916985211784, + "grad_norm": 0.9264475107192993, + "learning_rate": 0.00011433646086721494, + "loss": 0.181, + "step": 33280 + }, + { + "epoch": 1.2853778138152052, + "grad_norm": 0.2128441333770752, + "learning_rate": 0.00011431071984761317, + "loss": 0.1819, + "step": 33290 + }, + { + "epoch": 1.285763929109232, + "grad_norm": 0.5400950312614441, + "learning_rate": 0.0001142849788280114, + "loss": 0.4316, + "step": 33300 + }, + { + "epoch": 1.2861500444032588, + "grad_norm": 0.8033271431922913, + "learning_rate": 0.00011425923780840959, + "loss": 0.2146, + "step": 33310 + }, + { + "epoch": 1.2865361596972855, + "grad_norm": 2.012575149536133, + "learning_rate": 0.00011423349678880782, + "loss": 0.4335, + "step": 33320 + }, + { + "epoch": 1.2869222749913125, + "grad_norm": 0.7352376580238342, + "learning_rate": 0.00011420775576920602, + "loss": 0.2124, + "step": 33330 + }, + { + "epoch": 1.2873083902853393, + "grad_norm": 0.769036591053009, + "learning_rate": 0.00011418201474960423, + "loss": 0.3602, + "step": 33340 + }, + { + "epoch": 1.287694505579366, + "grad_norm": 0.250592976808548, + "learning_rate": 0.00011415627373000246, + "loss": 0.1692, + "step": 33350 + }, + { + "epoch": 1.2880806208733928, + "grad_norm": 2.43820858001709, + "learning_rate": 0.00011413053271040066, + "loss": 0.2777, + "step": 33360 + }, + { + "epoch": 1.2884667361674196, + "grad_norm": 1.3179954290390015, + "learning_rate": 0.00011410479169079889, + "loss": 0.1794, + "step": 33370 + }, + { + "epoch": 1.2888528514614463, + "grad_norm": 1.0040466785430908, + "learning_rate": 0.00011407905067119709, + "loss": 0.3037, + "step": 33380 + }, + { + "epoch": 1.2892389667554731, + "grad_norm": 5.296288013458252, + "learning_rate": 0.00011405330965159531, + "loss": 0.2904, + "step": 33390 + }, + { + "epoch": 1.2896250820495, + "grad_norm": 0.4267273247241974, + "learning_rate": 0.00011402756863199351, + "loss": 0.2263, + "step": 33400 + }, + { + "epoch": 1.2900111973435266, + "grad_norm": 0.8817713260650635, + "learning_rate": 0.00011400182761239173, + "loss": 0.2715, + "step": 33410 + }, + { + "epoch": 1.2903973126375536, + "grad_norm": 2.7891275882720947, + "learning_rate": 0.00011397608659278995, + "loss": 0.2781, + "step": 33420 + }, + { + "epoch": 1.2907834279315804, + "grad_norm": 0.3013952374458313, + "learning_rate": 0.00011395034557318815, + "loss": 0.2563, + "step": 33430 + }, + { + "epoch": 1.2911695432256072, + "grad_norm": 1.766413927078247, + "learning_rate": 0.00011392460455358638, + "loss": 0.1854, + "step": 33440 + }, + { + "epoch": 1.291555658519634, + "grad_norm": 0.25331103801727295, + "learning_rate": 0.00011389886353398458, + "loss": 0.1206, + "step": 33450 + }, + { + "epoch": 1.2919417738136607, + "grad_norm": 0.036400288343429565, + "learning_rate": 0.0001138731225143828, + "loss": 0.4707, + "step": 33460 + }, + { + "epoch": 1.2923278891076875, + "grad_norm": 1.5620888471603394, + "learning_rate": 0.00011384738149478102, + "loss": 0.3312, + "step": 33470 + }, + { + "epoch": 1.2927140044017142, + "grad_norm": 0.6670392155647278, + "learning_rate": 0.00011382164047517922, + "loss": 0.2341, + "step": 33480 + }, + { + "epoch": 1.2931001196957412, + "grad_norm": 2.3108737468719482, + "learning_rate": 0.00011379589945557745, + "loss": 0.3843, + "step": 33490 + }, + { + "epoch": 1.293486234989768, + "grad_norm": 0.8025147318840027, + "learning_rate": 0.00011377015843597565, + "loss": 0.1982, + "step": 33500 + }, + { + "epoch": 1.2938723502837948, + "grad_norm": 1.7835719585418701, + "learning_rate": 0.00011374441741637387, + "loss": 0.3285, + "step": 33510 + }, + { + "epoch": 1.2942584655778215, + "grad_norm": 2.041508913040161, + "learning_rate": 0.00011371867639677207, + "loss": 0.2044, + "step": 33520 + }, + { + "epoch": 1.2946445808718483, + "grad_norm": 1.103378415107727, + "learning_rate": 0.0001136929353771703, + "loss": 0.1682, + "step": 33530 + }, + { + "epoch": 1.295030696165875, + "grad_norm": 0.057376351207494736, + "learning_rate": 0.00011366719435756851, + "loss": 0.1642, + "step": 33540 + }, + { + "epoch": 1.2954168114599018, + "grad_norm": 0.6539410948753357, + "learning_rate": 0.00011364145333796671, + "loss": 0.1549, + "step": 33550 + }, + { + "epoch": 1.2958029267539288, + "grad_norm": 1.250543236732483, + "learning_rate": 0.00011361571231836494, + "loss": 0.3764, + "step": 33560 + }, + { + "epoch": 1.2961890420479556, + "grad_norm": 0.23697887361049652, + "learning_rate": 0.00011358997129876314, + "loss": 0.3999, + "step": 33570 + }, + { + "epoch": 1.2965751573419824, + "grad_norm": 0.9318505525588989, + "learning_rate": 0.00011356423027916137, + "loss": 0.4156, + "step": 33580 + }, + { + "epoch": 1.2969612726360091, + "grad_norm": 1.3910777568817139, + "learning_rate": 0.00011353848925955957, + "loss": 0.3455, + "step": 33590 + }, + { + "epoch": 1.297347387930036, + "grad_norm": 1.6764451265335083, + "learning_rate": 0.00011351274823995779, + "loss": 0.1884, + "step": 33600 + }, + { + "epoch": 1.2977335032240627, + "grad_norm": 0.9300051927566528, + "learning_rate": 0.000113487007220356, + "loss": 0.119, + "step": 33610 + }, + { + "epoch": 1.2981196185180894, + "grad_norm": 2.447462558746338, + "learning_rate": 0.00011346126620075422, + "loss": 0.4403, + "step": 33620 + }, + { + "epoch": 1.2985057338121164, + "grad_norm": 1.216407060623169, + "learning_rate": 0.00011343552518115243, + "loss": 0.2415, + "step": 33630 + }, + { + "epoch": 1.298891849106143, + "grad_norm": 2.968648910522461, + "learning_rate": 0.00011340978416155063, + "loss": 0.2899, + "step": 33640 + }, + { + "epoch": 1.29927796440017, + "grad_norm": 0.6649970412254333, + "learning_rate": 0.00011338404314194886, + "loss": 0.3809, + "step": 33650 + }, + { + "epoch": 1.2996640796941967, + "grad_norm": 1.7277917861938477, + "learning_rate": 0.00011335830212234709, + "loss": 0.3308, + "step": 33660 + }, + { + "epoch": 1.3000501949882235, + "grad_norm": 1.3269709348678589, + "learning_rate": 0.00011333256110274529, + "loss": 0.3682, + "step": 33670 + }, + { + "epoch": 1.3004363102822503, + "grad_norm": 0.20609407126903534, + "learning_rate": 0.0001133068200831435, + "loss": 0.1379, + "step": 33680 + }, + { + "epoch": 1.300822425576277, + "grad_norm": 0.6592215299606323, + "learning_rate": 0.00011328107906354171, + "loss": 0.2746, + "step": 33690 + }, + { + "epoch": 1.3012085408703038, + "grad_norm": 1.903635859489441, + "learning_rate": 0.00011325533804393993, + "loss": 0.4729, + "step": 33700 + }, + { + "epoch": 1.3015946561643306, + "grad_norm": 0.8432504534721375, + "learning_rate": 0.00011322959702433813, + "loss": 0.2835, + "step": 33710 + }, + { + "epoch": 1.3019807714583576, + "grad_norm": 0.9862542152404785, + "learning_rate": 0.00011320385600473635, + "loss": 0.1823, + "step": 33720 + }, + { + "epoch": 1.3023668867523843, + "grad_norm": 3.845738649368286, + "learning_rate": 0.00011317811498513458, + "loss": 0.2542, + "step": 33730 + }, + { + "epoch": 1.302753002046411, + "grad_norm": 0.6317747235298157, + "learning_rate": 0.00011315237396553278, + "loss": 0.22, + "step": 33740 + }, + { + "epoch": 1.3031391173404379, + "grad_norm": 2.5221354961395264, + "learning_rate": 0.000113126632945931, + "loss": 0.2253, + "step": 33750 + }, + { + "epoch": 1.3035252326344646, + "grad_norm": 1.3326247930526733, + "learning_rate": 0.0001131008919263292, + "loss": 0.2021, + "step": 33760 + }, + { + "epoch": 1.3039113479284914, + "grad_norm": 1.148047685623169, + "learning_rate": 0.00011307515090672742, + "loss": 0.3987, + "step": 33770 + }, + { + "epoch": 1.3042974632225182, + "grad_norm": 0.19721268117427826, + "learning_rate": 0.00011304940988712562, + "loss": 0.2642, + "step": 33780 + }, + { + "epoch": 1.3046835785165452, + "grad_norm": 1.4060617685317993, + "learning_rate": 0.00011302366886752385, + "loss": 0.2736, + "step": 33790 + }, + { + "epoch": 1.305069693810572, + "grad_norm": 1.0736548900604248, + "learning_rate": 0.00011299792784792207, + "loss": 0.2448, + "step": 33800 + }, + { + "epoch": 1.3054558091045987, + "grad_norm": 4.352476119995117, + "learning_rate": 0.00011297218682832027, + "loss": 0.383, + "step": 33810 + }, + { + "epoch": 1.3058419243986255, + "grad_norm": 0.2249228060245514, + "learning_rate": 0.0001129464458087185, + "loss": 0.14, + "step": 33820 + }, + { + "epoch": 1.3062280396926522, + "grad_norm": 0.4820781946182251, + "learning_rate": 0.0001129207047891167, + "loss": 0.248, + "step": 33830 + }, + { + "epoch": 1.306614154986679, + "grad_norm": 2.2983391284942627, + "learning_rate": 0.00011289496376951491, + "loss": 0.2608, + "step": 33840 + }, + { + "epoch": 1.3070002702807058, + "grad_norm": 1.3315671682357788, + "learning_rate": 0.00011286922274991314, + "loss": 0.1727, + "step": 33850 + }, + { + "epoch": 1.3073863855747327, + "grad_norm": 2.060299873352051, + "learning_rate": 0.00011284348173031134, + "loss": 0.3097, + "step": 33860 + }, + { + "epoch": 1.3077725008687593, + "grad_norm": 2.096285581588745, + "learning_rate": 0.00011281774071070957, + "loss": 0.2305, + "step": 33870 + }, + { + "epoch": 1.3081586161627863, + "grad_norm": 0.4997636675834656, + "learning_rate": 0.00011279199969110777, + "loss": 0.1993, + "step": 33880 + }, + { + "epoch": 1.308544731456813, + "grad_norm": 0.58636474609375, + "learning_rate": 0.00011276625867150599, + "loss": 0.1873, + "step": 33890 + }, + { + "epoch": 1.3089308467508398, + "grad_norm": 0.9128592610359192, + "learning_rate": 0.00011274051765190419, + "loss": 0.1885, + "step": 33900 + }, + { + "epoch": 1.3093169620448666, + "grad_norm": 2.228043794631958, + "learning_rate": 0.0001127147766323024, + "loss": 0.3649, + "step": 33910 + }, + { + "epoch": 1.3097030773388934, + "grad_norm": 1.069002389907837, + "learning_rate": 0.00011268903561270063, + "loss": 0.5454, + "step": 33920 + }, + { + "epoch": 1.3100891926329201, + "grad_norm": 0.6207597851753235, + "learning_rate": 0.00011266329459309883, + "loss": 0.2329, + "step": 33930 + }, + { + "epoch": 1.3104753079269469, + "grad_norm": 1.262247920036316, + "learning_rate": 0.00011263755357349706, + "loss": 0.3437, + "step": 33940 + }, + { + "epoch": 1.3108614232209739, + "grad_norm": 1.7429994344711304, + "learning_rate": 0.00011261181255389526, + "loss": 0.228, + "step": 33950 + }, + { + "epoch": 1.3112475385150006, + "grad_norm": 0.646900475025177, + "learning_rate": 0.00011258607153429349, + "loss": 0.3739, + "step": 33960 + }, + { + "epoch": 1.3116336538090274, + "grad_norm": 1.8228782415390015, + "learning_rate": 0.0001125603305146917, + "loss": 0.2325, + "step": 33970 + }, + { + "epoch": 1.3120197691030542, + "grad_norm": 3.539228916168213, + "learning_rate": 0.0001125345894950899, + "loss": 0.194, + "step": 33980 + }, + { + "epoch": 1.312405884397081, + "grad_norm": 1.2801135778427124, + "learning_rate": 0.00011250884847548813, + "loss": 0.3069, + "step": 33990 + }, + { + "epoch": 1.3127919996911077, + "grad_norm": 3.6265695095062256, + "learning_rate": 0.00011248310745588633, + "loss": 0.3113, + "step": 34000 + }, + { + "epoch": 1.3131781149851345, + "grad_norm": 0.07370063662528992, + "learning_rate": 0.00011245736643628455, + "loss": 0.1449, + "step": 34010 + }, + { + "epoch": 1.3135642302791615, + "grad_norm": 1.0295637845993042, + "learning_rate": 0.00011243162541668275, + "loss": 0.22, + "step": 34020 + }, + { + "epoch": 1.313950345573188, + "grad_norm": 0.8803662061691284, + "learning_rate": 0.00011240588439708098, + "loss": 0.1368, + "step": 34030 + }, + { + "epoch": 1.314336460867215, + "grad_norm": 1.6597707271575928, + "learning_rate": 0.00011238014337747919, + "loss": 0.3038, + "step": 34040 + }, + { + "epoch": 1.3147225761612418, + "grad_norm": 2.115492343902588, + "learning_rate": 0.00011235440235787739, + "loss": 0.1754, + "step": 34050 + }, + { + "epoch": 1.3151086914552685, + "grad_norm": 0.8143919706344604, + "learning_rate": 0.00011232866133827562, + "loss": 0.3764, + "step": 34060 + }, + { + "epoch": 1.3154948067492953, + "grad_norm": 0.14369767904281616, + "learning_rate": 0.00011230292031867382, + "loss": 0.1742, + "step": 34070 + }, + { + "epoch": 1.315880922043322, + "grad_norm": 1.0129845142364502, + "learning_rate": 0.00011227717929907205, + "loss": 0.1458, + "step": 34080 + }, + { + "epoch": 1.316267037337349, + "grad_norm": 2.7300291061401367, + "learning_rate": 0.00011225143827947025, + "loss": 0.3939, + "step": 34090 + }, + { + "epoch": 1.3166531526313756, + "grad_norm": 0.20205609500408173, + "learning_rate": 0.00011222569725986847, + "loss": 0.29, + "step": 34100 + }, + { + "epoch": 1.3170392679254026, + "grad_norm": 1.8928464651107788, + "learning_rate": 0.00011219995624026669, + "loss": 0.1742, + "step": 34110 + }, + { + "epoch": 1.3174253832194294, + "grad_norm": 0.2639687955379486, + "learning_rate": 0.00011217421522066488, + "loss": 0.1745, + "step": 34120 + }, + { + "epoch": 1.3178114985134561, + "grad_norm": 0.5906389355659485, + "learning_rate": 0.00011214847420106311, + "loss": 0.2134, + "step": 34130 + }, + { + "epoch": 1.318197613807483, + "grad_norm": 0.9190629720687866, + "learning_rate": 0.00011212273318146131, + "loss": 0.2547, + "step": 34140 + }, + { + "epoch": 1.3185837291015097, + "grad_norm": 0.5749151110649109, + "learning_rate": 0.00011209699216185954, + "loss": 0.1688, + "step": 34150 + }, + { + "epoch": 1.3189698443955364, + "grad_norm": 0.82295823097229, + "learning_rate": 0.00011207125114225777, + "loss": 0.2884, + "step": 34160 + }, + { + "epoch": 1.3193559596895632, + "grad_norm": 0.07816460728645325, + "learning_rate": 0.00011204551012265597, + "loss": 0.2418, + "step": 34170 + }, + { + "epoch": 1.3197420749835902, + "grad_norm": 0.6417407393455505, + "learning_rate": 0.00011201976910305418, + "loss": 0.2557, + "step": 34180 + }, + { + "epoch": 1.320128190277617, + "grad_norm": 6.093267440795898, + "learning_rate": 0.00011199402808345238, + "loss": 0.3088, + "step": 34190 + }, + { + "epoch": 1.3205143055716437, + "grad_norm": 1.8861887454986572, + "learning_rate": 0.0001119682870638506, + "loss": 0.2204, + "step": 34200 + }, + { + "epoch": 1.3209004208656705, + "grad_norm": 2.3272714614868164, + "learning_rate": 0.0001119425460442488, + "loss": 0.2236, + "step": 34210 + }, + { + "epoch": 1.3212865361596973, + "grad_norm": 0.9608810544013977, + "learning_rate": 0.00011191680502464703, + "loss": 0.1897, + "step": 34220 + }, + { + "epoch": 1.321672651453724, + "grad_norm": 1.2157350778579712, + "learning_rate": 0.00011189106400504526, + "loss": 0.1526, + "step": 34230 + }, + { + "epoch": 1.3220587667477508, + "grad_norm": 1.6684671640396118, + "learning_rate": 0.00011186532298544346, + "loss": 0.3394, + "step": 34240 + }, + { + "epoch": 1.3224448820417778, + "grad_norm": 2.0432374477386475, + "learning_rate": 0.00011183958196584167, + "loss": 0.2183, + "step": 34250 + }, + { + "epoch": 1.3228309973358043, + "grad_norm": 0.9436892867088318, + "learning_rate": 0.00011181384094623988, + "loss": 0.2947, + "step": 34260 + }, + { + "epoch": 1.3232171126298313, + "grad_norm": 0.23260092735290527, + "learning_rate": 0.0001117880999266381, + "loss": 0.114, + "step": 34270 + }, + { + "epoch": 1.323603227923858, + "grad_norm": 1.2291594743728638, + "learning_rate": 0.0001117623589070363, + "loss": 0.3145, + "step": 34280 + }, + { + "epoch": 1.3239893432178849, + "grad_norm": 0.41411107778549194, + "learning_rate": 0.00011173661788743452, + "loss": 0.2937, + "step": 34290 + }, + { + "epoch": 1.3243754585119116, + "grad_norm": 2.354405164718628, + "learning_rate": 0.00011171087686783275, + "loss": 0.3933, + "step": 34300 + }, + { + "epoch": 1.3247615738059384, + "grad_norm": 2.6997978687286377, + "learning_rate": 0.00011168513584823095, + "loss": 0.1494, + "step": 34310 + }, + { + "epoch": 1.3251476890999654, + "grad_norm": 2.8430919647216797, + "learning_rate": 0.00011165939482862916, + "loss": 0.2869, + "step": 34320 + }, + { + "epoch": 1.325533804393992, + "grad_norm": 1.1737356185913086, + "learning_rate": 0.00011163365380902738, + "loss": 0.2792, + "step": 34330 + }, + { + "epoch": 1.325919919688019, + "grad_norm": 4.123973846435547, + "learning_rate": 0.00011160791278942559, + "loss": 0.5211, + "step": 34340 + }, + { + "epoch": 1.3263060349820457, + "grad_norm": 0.8862038850784302, + "learning_rate": 0.00011158217176982382, + "loss": 0.2976, + "step": 34350 + }, + { + "epoch": 1.3266921502760725, + "grad_norm": 1.8690590858459473, + "learning_rate": 0.00011155643075022202, + "loss": 0.2485, + "step": 34360 + }, + { + "epoch": 1.3270782655700992, + "grad_norm": 2.885589599609375, + "learning_rate": 0.00011153068973062024, + "loss": 0.3084, + "step": 34370 + }, + { + "epoch": 1.327464380864126, + "grad_norm": 0.9898788928985596, + "learning_rate": 0.00011150494871101844, + "loss": 0.261, + "step": 34380 + }, + { + "epoch": 1.3278504961581528, + "grad_norm": 0.6879653930664062, + "learning_rate": 0.00011147920769141667, + "loss": 0.2082, + "step": 34390 + }, + { + "epoch": 1.3282366114521795, + "grad_norm": 1.2619003057479858, + "learning_rate": 0.00011145346667181487, + "loss": 0.2402, + "step": 34400 + }, + { + "epoch": 1.3286227267462065, + "grad_norm": 1.1212007999420166, + "learning_rate": 0.00011142772565221308, + "loss": 0.3525, + "step": 34410 + }, + { + "epoch": 1.3290088420402333, + "grad_norm": 1.8431956768035889, + "learning_rate": 0.00011140198463261131, + "loss": 0.2212, + "step": 34420 + }, + { + "epoch": 1.32939495733426, + "grad_norm": 0.6185423731803894, + "learning_rate": 0.00011137624361300951, + "loss": 0.2455, + "step": 34430 + }, + { + "epoch": 1.3297810726282868, + "grad_norm": 2.3791301250457764, + "learning_rate": 0.00011135050259340774, + "loss": 0.1763, + "step": 34440 + }, + { + "epoch": 1.3301671879223136, + "grad_norm": 0.4928603768348694, + "learning_rate": 0.00011132476157380594, + "loss": 0.2381, + "step": 34450 + }, + { + "epoch": 1.3305533032163404, + "grad_norm": 1.5636029243469238, + "learning_rate": 0.00011129902055420416, + "loss": 0.1368, + "step": 34460 + }, + { + "epoch": 1.3309394185103671, + "grad_norm": 0.9425283074378967, + "learning_rate": 0.00011127327953460238, + "loss": 0.2825, + "step": 34470 + }, + { + "epoch": 1.3313255338043941, + "grad_norm": 1.2257115840911865, + "learning_rate": 0.00011124753851500058, + "loss": 0.2547, + "step": 34480 + }, + { + "epoch": 1.3317116490984207, + "grad_norm": 0.9416170716285706, + "learning_rate": 0.0001112217974953988, + "loss": 0.2766, + "step": 34490 + }, + { + "epoch": 1.3320977643924476, + "grad_norm": 0.5123847126960754, + "learning_rate": 0.000111196056475797, + "loss": 0.4733, + "step": 34500 + }, + { + "epoch": 1.3324838796864744, + "grad_norm": 1.5581384897232056, + "learning_rate": 0.00011117031545619523, + "loss": 0.1597, + "step": 34510 + }, + { + "epoch": 1.3328699949805012, + "grad_norm": 2.377333879470825, + "learning_rate": 0.00011114457443659343, + "loss": 0.209, + "step": 34520 + }, + { + "epoch": 1.333256110274528, + "grad_norm": 1.7840913534164429, + "learning_rate": 0.00011111883341699166, + "loss": 0.1759, + "step": 34530 + }, + { + "epoch": 1.3336422255685547, + "grad_norm": 1.1825993061065674, + "learning_rate": 0.00011109309239738987, + "loss": 0.2464, + "step": 34540 + }, + { + "epoch": 1.3340283408625815, + "grad_norm": 1.8859659433364868, + "learning_rate": 0.00011106735137778807, + "loss": 0.3539, + "step": 34550 + }, + { + "epoch": 1.3344144561566083, + "grad_norm": 1.9698175191879272, + "learning_rate": 0.0001110416103581863, + "loss": 0.3301, + "step": 34560 + }, + { + "epoch": 1.3348005714506352, + "grad_norm": 0.7649385333061218, + "learning_rate": 0.0001110158693385845, + "loss": 0.232, + "step": 34570 + }, + { + "epoch": 1.335186686744662, + "grad_norm": 0.56386399269104, + "learning_rate": 0.00011099012831898272, + "loss": 0.3425, + "step": 34580 + }, + { + "epoch": 1.3355728020386888, + "grad_norm": 2.956003189086914, + "learning_rate": 0.00011096438729938092, + "loss": 0.1518, + "step": 34590 + }, + { + "epoch": 1.3359589173327155, + "grad_norm": 2.612029552459717, + "learning_rate": 0.00011093864627977915, + "loss": 0.2765, + "step": 34600 + }, + { + "epoch": 1.3363450326267423, + "grad_norm": 0.9674397706985474, + "learning_rate": 0.00011091290526017736, + "loss": 0.303, + "step": 34610 + }, + { + "epoch": 1.336731147920769, + "grad_norm": 0.9578921794891357, + "learning_rate": 0.00011088716424057556, + "loss": 0.1405, + "step": 34620 + }, + { + "epoch": 1.3371172632147958, + "grad_norm": 2.168065071105957, + "learning_rate": 0.00011086142322097379, + "loss": 0.1914, + "step": 34630 + }, + { + "epoch": 1.3375033785088228, + "grad_norm": 1.3166526556015015, + "learning_rate": 0.00011083568220137199, + "loss": 0.4134, + "step": 34640 + }, + { + "epoch": 1.3378894938028496, + "grad_norm": 0.9082283973693848, + "learning_rate": 0.00011080994118177022, + "loss": 0.2693, + "step": 34650 + }, + { + "epoch": 1.3382756090968764, + "grad_norm": 2.203007698059082, + "learning_rate": 0.00011078420016216844, + "loss": 0.1847, + "step": 34660 + }, + { + "epoch": 1.3386617243909031, + "grad_norm": 0.8101674914360046, + "learning_rate": 0.00011075845914256664, + "loss": 0.3111, + "step": 34670 + }, + { + "epoch": 1.33904783968493, + "grad_norm": 1.9545695781707764, + "learning_rate": 0.00011073271812296486, + "loss": 0.3761, + "step": 34680 + }, + { + "epoch": 1.3394339549789567, + "grad_norm": 1.547581672668457, + "learning_rate": 0.00011070697710336306, + "loss": 0.2374, + "step": 34690 + }, + { + "epoch": 1.3398200702729834, + "grad_norm": 3.3519034385681152, + "learning_rate": 0.00011068123608376128, + "loss": 0.1957, + "step": 34700 + }, + { + "epoch": 1.3402061855670104, + "grad_norm": 1.5508599281311035, + "learning_rate": 0.00011065549506415948, + "loss": 0.4171, + "step": 34710 + }, + { + "epoch": 1.340592300861037, + "grad_norm": 1.8547546863555908, + "learning_rate": 0.00011062975404455771, + "loss": 0.1872, + "step": 34720 + }, + { + "epoch": 1.340978416155064, + "grad_norm": 1.4600756168365479, + "learning_rate": 0.00011060401302495594, + "loss": 0.3515, + "step": 34730 + }, + { + "epoch": 1.3413645314490907, + "grad_norm": 0.05774044618010521, + "learning_rate": 0.00011057827200535414, + "loss": 0.1604, + "step": 34740 + }, + { + "epoch": 1.3417506467431175, + "grad_norm": 2.8793342113494873, + "learning_rate": 0.00011055253098575235, + "loss": 0.3095, + "step": 34750 + }, + { + "epoch": 1.3421367620371443, + "grad_norm": 2.241042375564575, + "learning_rate": 0.00011052678996615055, + "loss": 0.2511, + "step": 34760 + }, + { + "epoch": 1.342522877331171, + "grad_norm": 1.9320632219314575, + "learning_rate": 0.00011050104894654878, + "loss": 0.4493, + "step": 34770 + }, + { + "epoch": 1.3429089926251978, + "grad_norm": 1.6483882665634155, + "learning_rate": 0.000110475307926947, + "loss": 0.217, + "step": 34780 + }, + { + "epoch": 1.3432951079192246, + "grad_norm": 0.9635765552520752, + "learning_rate": 0.0001104495669073452, + "loss": 0.5458, + "step": 34790 + }, + { + "epoch": 1.3436812232132516, + "grad_norm": 1.2436567544937134, + "learning_rate": 0.00011042382588774343, + "loss": 0.2857, + "step": 34800 + }, + { + "epoch": 1.3440673385072783, + "grad_norm": 2.8082425594329834, + "learning_rate": 0.00011039808486814163, + "loss": 0.3439, + "step": 34810 + }, + { + "epoch": 1.344453453801305, + "grad_norm": 1.0430901050567627, + "learning_rate": 0.00011037234384853984, + "loss": 0.1404, + "step": 34820 + }, + { + "epoch": 1.3448395690953319, + "grad_norm": 1.7387149333953857, + "learning_rate": 0.00011034660282893806, + "loss": 0.395, + "step": 34830 + }, + { + "epoch": 1.3452256843893586, + "grad_norm": 1.2713748216629028, + "learning_rate": 0.00011032086180933627, + "loss": 0.29, + "step": 34840 + }, + { + "epoch": 1.3456117996833854, + "grad_norm": 0.26068204641342163, + "learning_rate": 0.0001102951207897345, + "loss": 0.1814, + "step": 34850 + }, + { + "epoch": 1.3459979149774122, + "grad_norm": 2.4163243770599365, + "learning_rate": 0.0001102693797701327, + "loss": 0.2445, + "step": 34860 + }, + { + "epoch": 1.3463840302714392, + "grad_norm": 2.2439687252044678, + "learning_rate": 0.00011024363875053092, + "loss": 0.3338, + "step": 34870 + }, + { + "epoch": 1.346770145565466, + "grad_norm": 0.2822403609752655, + "learning_rate": 0.00011021789773092912, + "loss": 0.1648, + "step": 34880 + }, + { + "epoch": 1.3471562608594927, + "grad_norm": 0.07319017499685287, + "learning_rate": 0.00011019215671132734, + "loss": 0.107, + "step": 34890 + }, + { + "epoch": 1.3475423761535195, + "grad_norm": 0.9809044003486633, + "learning_rate": 0.00011016641569172555, + "loss": 0.256, + "step": 34900 + }, + { + "epoch": 1.3479284914475462, + "grad_norm": 0.5016226768493652, + "learning_rate": 0.00011014067467212376, + "loss": 0.3229, + "step": 34910 + }, + { + "epoch": 1.348314606741573, + "grad_norm": 1.3026005029678345, + "learning_rate": 0.00011011493365252199, + "loss": 0.2145, + "step": 34920 + }, + { + "epoch": 1.3487007220355998, + "grad_norm": 1.0752215385437012, + "learning_rate": 0.00011008919263292019, + "loss": 0.2355, + "step": 34930 + }, + { + "epoch": 1.3490868373296268, + "grad_norm": 2.2703003883361816, + "learning_rate": 0.00011006345161331842, + "loss": 0.2079, + "step": 34940 + }, + { + "epoch": 1.3494729526236533, + "grad_norm": 1.1323810815811157, + "learning_rate": 0.00011003771059371662, + "loss": 0.1015, + "step": 34950 + }, + { + "epoch": 1.3498590679176803, + "grad_norm": 0.10813555121421814, + "learning_rate": 0.00011001196957411484, + "loss": 0.4214, + "step": 34960 + }, + { + "epoch": 1.350245183211707, + "grad_norm": 0.07815568149089813, + "learning_rate": 0.00010998622855451306, + "loss": 0.1211, + "step": 34970 + }, + { + "epoch": 1.3506312985057338, + "grad_norm": 0.6748234629631042, + "learning_rate": 0.00010996048753491126, + "loss": 0.3508, + "step": 34980 + }, + { + "epoch": 1.3510174137997606, + "grad_norm": 1.8556997776031494, + "learning_rate": 0.00010993474651530948, + "loss": 0.2268, + "step": 34990 + }, + { + "epoch": 1.3514035290937874, + "grad_norm": 0.8696061372756958, + "learning_rate": 0.00010990900549570768, + "loss": 0.4321, + "step": 35000 + }, + { + "epoch": 1.3517896443878141, + "grad_norm": 0.42442765831947327, + "learning_rate": 0.00010988326447610591, + "loss": 0.1944, + "step": 35010 + }, + { + "epoch": 1.352175759681841, + "grad_norm": 1.0474554300308228, + "learning_rate": 0.00010985752345650411, + "loss": 0.1342, + "step": 35020 + }, + { + "epoch": 1.3525618749758679, + "grad_norm": 0.607037365436554, + "learning_rate": 0.00010983178243690234, + "loss": 0.2965, + "step": 35030 + }, + { + "epoch": 1.3529479902698947, + "grad_norm": 1.8160990476608276, + "learning_rate": 0.00010980604141730055, + "loss": 0.3192, + "step": 35040 + }, + { + "epoch": 1.3533341055639214, + "grad_norm": 2.0026509761810303, + "learning_rate": 0.00010978030039769875, + "loss": 0.3054, + "step": 35050 + }, + { + "epoch": 1.3537202208579482, + "grad_norm": 0.9203600883483887, + "learning_rate": 0.00010975455937809698, + "loss": 0.253, + "step": 35060 + }, + { + "epoch": 1.354106336151975, + "grad_norm": 0.33198195695877075, + "learning_rate": 0.00010972881835849518, + "loss": 0.3885, + "step": 35070 + }, + { + "epoch": 1.3544924514460017, + "grad_norm": 0.3201223611831665, + "learning_rate": 0.0001097030773388934, + "loss": 0.3029, + "step": 35080 + }, + { + "epoch": 1.3548785667400285, + "grad_norm": 1.2589943408966064, + "learning_rate": 0.0001096773363192916, + "loss": 0.4243, + "step": 35090 + }, + { + "epoch": 1.3552646820340555, + "grad_norm": 1.5106219053268433, + "learning_rate": 0.00010965159529968983, + "loss": 0.2585, + "step": 35100 + }, + { + "epoch": 1.3556507973280822, + "grad_norm": 1.429799199104309, + "learning_rate": 0.00010962585428008804, + "loss": 0.1961, + "step": 35110 + }, + { + "epoch": 1.356036912622109, + "grad_norm": 2.1211297512054443, + "learning_rate": 0.00010960011326048624, + "loss": 0.4057, + "step": 35120 + }, + { + "epoch": 1.3564230279161358, + "grad_norm": 2.5154731273651123, + "learning_rate": 0.00010957437224088447, + "loss": 0.3787, + "step": 35130 + }, + { + "epoch": 1.3568091432101625, + "grad_norm": 0.4914834201335907, + "learning_rate": 0.00010954863122128267, + "loss": 0.234, + "step": 35140 + }, + { + "epoch": 1.3571952585041893, + "grad_norm": 0.26685893535614014, + "learning_rate": 0.0001095228902016809, + "loss": 0.2841, + "step": 35150 + }, + { + "epoch": 1.357581373798216, + "grad_norm": 0.15462155640125275, + "learning_rate": 0.00010949714918207912, + "loss": 0.2269, + "step": 35160 + }, + { + "epoch": 1.357967489092243, + "grad_norm": 1.3887063264846802, + "learning_rate": 0.00010947140816247732, + "loss": 0.3455, + "step": 35170 + }, + { + "epoch": 1.3583536043862696, + "grad_norm": 0.786374032497406, + "learning_rate": 0.00010944566714287554, + "loss": 0.2897, + "step": 35180 + }, + { + "epoch": 1.3587397196802966, + "grad_norm": 1.100475549697876, + "learning_rate": 0.00010941992612327374, + "loss": 0.2892, + "step": 35190 + }, + { + "epoch": 1.3591258349743234, + "grad_norm": 0.7676102519035339, + "learning_rate": 0.00010939418510367196, + "loss": 0.1942, + "step": 35200 + }, + { + "epoch": 1.3595119502683501, + "grad_norm": 0.33462053537368774, + "learning_rate": 0.00010936844408407016, + "loss": 0.2872, + "step": 35210 + }, + { + "epoch": 1.359898065562377, + "grad_norm": 0.9294387698173523, + "learning_rate": 0.00010934270306446839, + "loss": 0.2617, + "step": 35220 + }, + { + "epoch": 1.3602841808564037, + "grad_norm": 0.3169979453086853, + "learning_rate": 0.00010931696204486662, + "loss": 0.2942, + "step": 35230 + }, + { + "epoch": 1.3606702961504304, + "grad_norm": 2.1339616775512695, + "learning_rate": 0.00010929122102526482, + "loss": 0.4448, + "step": 35240 + }, + { + "epoch": 1.3610564114444572, + "grad_norm": 0.9430062770843506, + "learning_rate": 0.00010926548000566303, + "loss": 0.2051, + "step": 35250 + }, + { + "epoch": 1.3614425267384842, + "grad_norm": 3.1187360286712646, + "learning_rate": 0.00010923973898606123, + "loss": 0.2274, + "step": 35260 + }, + { + "epoch": 1.361828642032511, + "grad_norm": 1.4727579355239868, + "learning_rate": 0.00010921399796645946, + "loss": 0.3757, + "step": 35270 + }, + { + "epoch": 1.3622147573265377, + "grad_norm": 2.157560348510742, + "learning_rate": 0.00010918825694685768, + "loss": 0.3096, + "step": 35280 + }, + { + "epoch": 1.3626008726205645, + "grad_norm": 0.33457377552986145, + "learning_rate": 0.00010916251592725588, + "loss": 0.1489, + "step": 35290 + }, + { + "epoch": 1.3629869879145913, + "grad_norm": 0.9005904197692871, + "learning_rate": 0.00010913677490765411, + "loss": 0.1826, + "step": 35300 + }, + { + "epoch": 1.363373103208618, + "grad_norm": 2.1222829818725586, + "learning_rate": 0.00010911103388805231, + "loss": 0.1965, + "step": 35310 + }, + { + "epoch": 1.3637592185026448, + "grad_norm": 1.3881357908248901, + "learning_rate": 0.00010908529286845052, + "loss": 0.1791, + "step": 35320 + }, + { + "epoch": 1.3641453337966718, + "grad_norm": 1.7574503421783447, + "learning_rate": 0.00010905955184884872, + "loss": 0.3316, + "step": 35330 + }, + { + "epoch": 1.3645314490906983, + "grad_norm": 0.1967727392911911, + "learning_rate": 0.00010903381082924695, + "loss": 0.2331, + "step": 35340 + }, + { + "epoch": 1.3649175643847253, + "grad_norm": 0.8974360823631287, + "learning_rate": 0.00010900806980964518, + "loss": 0.2589, + "step": 35350 + }, + { + "epoch": 1.365303679678752, + "grad_norm": 2.0996744632720947, + "learning_rate": 0.00010898232879004338, + "loss": 0.3663, + "step": 35360 + }, + { + "epoch": 1.3656897949727789, + "grad_norm": 0.5678316354751587, + "learning_rate": 0.0001089565877704416, + "loss": 0.1729, + "step": 35370 + }, + { + "epoch": 1.3660759102668056, + "grad_norm": 2.3381874561309814, + "learning_rate": 0.0001089308467508398, + "loss": 0.1615, + "step": 35380 + }, + { + "epoch": 1.3664620255608324, + "grad_norm": 1.0276836156845093, + "learning_rate": 0.00010890510573123802, + "loss": 0.3359, + "step": 35390 + }, + { + "epoch": 1.3668481408548594, + "grad_norm": 2.4374940395355225, + "learning_rate": 0.00010887936471163622, + "loss": 0.2435, + "step": 35400 + }, + { + "epoch": 1.367234256148886, + "grad_norm": 0.45221665501594543, + "learning_rate": 0.00010885362369203444, + "loss": 0.2555, + "step": 35410 + }, + { + "epoch": 1.367620371442913, + "grad_norm": 2.608090400695801, + "learning_rate": 0.00010882788267243267, + "loss": 0.2465, + "step": 35420 + }, + { + "epoch": 1.3680064867369397, + "grad_norm": 1.4186642169952393, + "learning_rate": 0.00010880214165283087, + "loss": 0.1674, + "step": 35430 + }, + { + "epoch": 1.3683926020309665, + "grad_norm": 0.659479022026062, + "learning_rate": 0.0001087764006332291, + "loss": 0.2926, + "step": 35440 + }, + { + "epoch": 1.3687787173249932, + "grad_norm": 0.9219567179679871, + "learning_rate": 0.0001087506596136273, + "loss": 0.2001, + "step": 35450 + }, + { + "epoch": 1.36916483261902, + "grad_norm": 0.8070804476737976, + "learning_rate": 0.00010872491859402551, + "loss": 0.2178, + "step": 35460 + }, + { + "epoch": 1.3695509479130468, + "grad_norm": 2.9981069564819336, + "learning_rate": 0.00010869917757442374, + "loss": 0.3079, + "step": 35470 + }, + { + "epoch": 1.3699370632070735, + "grad_norm": 0.7891242504119873, + "learning_rate": 0.00010867343655482194, + "loss": 0.2765, + "step": 35480 + }, + { + "epoch": 1.3703231785011005, + "grad_norm": 1.448637843132019, + "learning_rate": 0.00010864769553522016, + "loss": 0.3521, + "step": 35490 + }, + { + "epoch": 1.3707092937951273, + "grad_norm": 0.07628043740987778, + "learning_rate": 0.00010862195451561836, + "loss": 0.2083, + "step": 35500 + }, + { + "epoch": 1.371095409089154, + "grad_norm": 0.7549735307693481, + "learning_rate": 0.00010859621349601659, + "loss": 0.2536, + "step": 35510 + }, + { + "epoch": 1.3714815243831808, + "grad_norm": 1.3548041582107544, + "learning_rate": 0.00010857047247641479, + "loss": 0.251, + "step": 35520 + }, + { + "epoch": 1.3718676396772076, + "grad_norm": 0.530010998249054, + "learning_rate": 0.000108544731456813, + "loss": 0.1917, + "step": 35530 + }, + { + "epoch": 1.3722537549712344, + "grad_norm": 0.4148992896080017, + "learning_rate": 0.00010851899043721123, + "loss": 0.335, + "step": 35540 + }, + { + "epoch": 1.3726398702652611, + "grad_norm": 1.5118776559829712, + "learning_rate": 0.00010849324941760943, + "loss": 0.2159, + "step": 35550 + }, + { + "epoch": 1.3730259855592881, + "grad_norm": 1.036889910697937, + "learning_rate": 0.00010846750839800766, + "loss": 0.2975, + "step": 35560 + }, + { + "epoch": 1.3734121008533147, + "grad_norm": 1.724263072013855, + "learning_rate": 0.00010844176737840586, + "loss": 0.1476, + "step": 35570 + }, + { + "epoch": 1.3737982161473417, + "grad_norm": 1.599007487297058, + "learning_rate": 0.00010841602635880408, + "loss": 0.2539, + "step": 35580 + }, + { + "epoch": 1.3741843314413684, + "grad_norm": 2.9119279384613037, + "learning_rate": 0.00010839028533920228, + "loss": 0.2688, + "step": 35590 + }, + { + "epoch": 1.3745704467353952, + "grad_norm": 1.8647874593734741, + "learning_rate": 0.00010836454431960051, + "loss": 0.4158, + "step": 35600 + }, + { + "epoch": 1.374956562029422, + "grad_norm": 3.925290822982788, + "learning_rate": 0.00010833880329999872, + "loss": 0.3333, + "step": 35610 + }, + { + "epoch": 1.3753426773234487, + "grad_norm": 0.7124634385108948, + "learning_rate": 0.00010831306228039692, + "loss": 0.1069, + "step": 35620 + }, + { + "epoch": 1.3757287926174757, + "grad_norm": 1.303579330444336, + "learning_rate": 0.00010828732126079515, + "loss": 0.2898, + "step": 35630 + }, + { + "epoch": 1.3761149079115023, + "grad_norm": 3.921804189682007, + "learning_rate": 0.00010826158024119335, + "loss": 0.4212, + "step": 35640 + }, + { + "epoch": 1.3765010232055293, + "grad_norm": 1.3194564580917358, + "learning_rate": 0.00010823583922159158, + "loss": 0.2771, + "step": 35650 + }, + { + "epoch": 1.376887138499556, + "grad_norm": 1.4237637519836426, + "learning_rate": 0.00010821009820198979, + "loss": 0.2463, + "step": 35660 + }, + { + "epoch": 1.3772732537935828, + "grad_norm": 1.8165888786315918, + "learning_rate": 0.000108184357182388, + "loss": 0.291, + "step": 35670 + }, + { + "epoch": 1.3776593690876096, + "grad_norm": 1.1056426763534546, + "learning_rate": 0.00010815861616278622, + "loss": 0.2525, + "step": 35680 + }, + { + "epoch": 1.3780454843816363, + "grad_norm": 1.483189582824707, + "learning_rate": 0.00010813287514318442, + "loss": 0.1569, + "step": 35690 + }, + { + "epoch": 1.378431599675663, + "grad_norm": 1.0666841268539429, + "learning_rate": 0.00010810713412358264, + "loss": 0.235, + "step": 35700 + }, + { + "epoch": 1.3788177149696899, + "grad_norm": 1.0299845933914185, + "learning_rate": 0.00010808139310398084, + "loss": 0.3892, + "step": 35710 + }, + { + "epoch": 1.3792038302637168, + "grad_norm": 2.3474409580230713, + "learning_rate": 0.00010805565208437907, + "loss": 0.3417, + "step": 35720 + }, + { + "epoch": 1.3795899455577436, + "grad_norm": 1.7456315755844116, + "learning_rate": 0.0001080299110647773, + "loss": 0.2538, + "step": 35730 + }, + { + "epoch": 1.3799760608517704, + "grad_norm": 2.866103410720825, + "learning_rate": 0.0001080041700451755, + "loss": 0.1619, + "step": 35740 + }, + { + "epoch": 1.3803621761457971, + "grad_norm": 0.29136407375335693, + "learning_rate": 0.00010797842902557371, + "loss": 0.2692, + "step": 35750 + }, + { + "epoch": 1.380748291439824, + "grad_norm": 0.8046161532402039, + "learning_rate": 0.00010795268800597191, + "loss": 0.1575, + "step": 35760 + }, + { + "epoch": 1.3811344067338507, + "grad_norm": 0.6451787352561951, + "learning_rate": 0.00010792694698637014, + "loss": 0.4914, + "step": 35770 + }, + { + "epoch": 1.3815205220278775, + "grad_norm": 0.7289161086082458, + "learning_rate": 0.00010790120596676836, + "loss": 0.1895, + "step": 35780 + }, + { + "epoch": 1.3819066373219044, + "grad_norm": 0.8300430178642273, + "learning_rate": 0.00010787546494716656, + "loss": 0.3663, + "step": 35790 + }, + { + "epoch": 1.382292752615931, + "grad_norm": 0.17713364958763123, + "learning_rate": 0.00010784972392756479, + "loss": 0.3189, + "step": 35800 + }, + { + "epoch": 1.382678867909958, + "grad_norm": 0.903222918510437, + "learning_rate": 0.00010782398290796299, + "loss": 0.1577, + "step": 35810 + }, + { + "epoch": 1.3830649832039847, + "grad_norm": 0.08617932349443436, + "learning_rate": 0.0001077982418883612, + "loss": 0.2872, + "step": 35820 + }, + { + "epoch": 1.3834510984980115, + "grad_norm": 1.9590895175933838, + "learning_rate": 0.0001077725008687594, + "loss": 0.2907, + "step": 35830 + }, + { + "epoch": 1.3838372137920383, + "grad_norm": 1.2515161037445068, + "learning_rate": 0.00010774675984915763, + "loss": 0.177, + "step": 35840 + }, + { + "epoch": 1.384223329086065, + "grad_norm": 1.6171292066574097, + "learning_rate": 0.00010772101882955586, + "loss": 0.2321, + "step": 35850 + }, + { + "epoch": 1.3846094443800918, + "grad_norm": 0.13681405782699585, + "learning_rate": 0.00010769527780995406, + "loss": 0.247, + "step": 35860 + }, + { + "epoch": 1.3849955596741186, + "grad_norm": 1.1949968338012695, + "learning_rate": 0.00010766953679035228, + "loss": 0.2978, + "step": 35870 + }, + { + "epoch": 1.3853816749681456, + "grad_norm": 0.17001692950725555, + "learning_rate": 0.00010764379577075048, + "loss": 0.3241, + "step": 35880 + }, + { + "epoch": 1.3857677902621723, + "grad_norm": 0.8227952122688293, + "learning_rate": 0.0001076180547511487, + "loss": 0.3499, + "step": 35890 + }, + { + "epoch": 1.386153905556199, + "grad_norm": 1.4185482263565063, + "learning_rate": 0.0001075923137315469, + "loss": 0.3109, + "step": 35900 + }, + { + "epoch": 1.3865400208502259, + "grad_norm": 0.9533351063728333, + "learning_rate": 0.00010756657271194512, + "loss": 0.24, + "step": 35910 + }, + { + "epoch": 1.3869261361442526, + "grad_norm": 1.077789306640625, + "learning_rate": 0.00010754083169234335, + "loss": 0.2662, + "step": 35920 + }, + { + "epoch": 1.3873122514382794, + "grad_norm": 1.3528363704681396, + "learning_rate": 0.00010751509067274155, + "loss": 0.1623, + "step": 35930 + }, + { + "epoch": 1.3876983667323062, + "grad_norm": 0.25122806429862976, + "learning_rate": 0.00010748934965313978, + "loss": 0.1639, + "step": 35940 + }, + { + "epoch": 1.3880844820263332, + "grad_norm": 0.9446159601211548, + "learning_rate": 0.00010746360863353798, + "loss": 0.2035, + "step": 35950 + }, + { + "epoch": 1.38847059732036, + "grad_norm": 1.2258719205856323, + "learning_rate": 0.00010743786761393619, + "loss": 0.4247, + "step": 35960 + }, + { + "epoch": 1.3888567126143867, + "grad_norm": 3.117729663848877, + "learning_rate": 0.00010741212659433442, + "loss": 0.2569, + "step": 35970 + }, + { + "epoch": 1.3892428279084135, + "grad_norm": 0.6333123445510864, + "learning_rate": 0.00010738638557473262, + "loss": 0.1843, + "step": 35980 + }, + { + "epoch": 1.3896289432024402, + "grad_norm": 1.49360191822052, + "learning_rate": 0.00010736064455513084, + "loss": 0.2318, + "step": 35990 + }, + { + "epoch": 1.390015058496467, + "grad_norm": 3.9082753658294678, + "learning_rate": 0.00010733490353552904, + "loss": 0.2037, + "step": 36000 + }, + { + "epoch": 1.3904011737904938, + "grad_norm": 0.5687323808670044, + "learning_rate": 0.00010730916251592727, + "loss": 0.0871, + "step": 36010 + }, + { + "epoch": 1.3907872890845208, + "grad_norm": 0.3480868339538574, + "learning_rate": 0.00010728342149632547, + "loss": 0.4009, + "step": 36020 + }, + { + "epoch": 1.3911734043785473, + "grad_norm": 1.325042486190796, + "learning_rate": 0.00010725768047672368, + "loss": 0.5143, + "step": 36030 + }, + { + "epoch": 1.3915595196725743, + "grad_norm": 2.114786386489868, + "learning_rate": 0.00010723193945712191, + "loss": 0.2944, + "step": 36040 + }, + { + "epoch": 1.391945634966601, + "grad_norm": 1.716272234916687, + "learning_rate": 0.00010720619843752011, + "loss": 0.3793, + "step": 36050 + }, + { + "epoch": 1.3923317502606278, + "grad_norm": 2.057535171508789, + "learning_rate": 0.00010718045741791834, + "loss": 0.199, + "step": 36060 + }, + { + "epoch": 1.3927178655546546, + "grad_norm": 2.371248483657837, + "learning_rate": 0.00010715471639831653, + "loss": 0.3276, + "step": 36070 + }, + { + "epoch": 1.3931039808486814, + "grad_norm": 0.1784186065196991, + "learning_rate": 0.00010712897537871476, + "loss": 0.0922, + "step": 36080 + }, + { + "epoch": 1.3934900961427081, + "grad_norm": 0.7923040390014648, + "learning_rate": 0.00010710323435911298, + "loss": 0.1599, + "step": 36090 + }, + { + "epoch": 1.393876211436735, + "grad_norm": 1.9013831615447998, + "learning_rate": 0.00010707749333951117, + "loss": 0.4997, + "step": 36100 + }, + { + "epoch": 1.394262326730762, + "grad_norm": 1.8659415245056152, + "learning_rate": 0.0001070517523199094, + "loss": 0.1851, + "step": 36110 + }, + { + "epoch": 1.3946484420247887, + "grad_norm": 2.5775375366210938, + "learning_rate": 0.0001070260113003076, + "loss": 0.2384, + "step": 36120 + }, + { + "epoch": 1.3950345573188154, + "grad_norm": 0.21943879127502441, + "learning_rate": 0.00010700027028070583, + "loss": 0.3996, + "step": 36130 + }, + { + "epoch": 1.3954206726128422, + "grad_norm": 1.1734743118286133, + "learning_rate": 0.00010697452926110403, + "loss": 0.1737, + "step": 36140 + }, + { + "epoch": 1.395806787906869, + "grad_norm": 0.697695791721344, + "learning_rate": 0.00010694878824150225, + "loss": 0.2778, + "step": 36150 + }, + { + "epoch": 1.3961929032008957, + "grad_norm": 3.2881579399108887, + "learning_rate": 0.00010692304722190047, + "loss": 0.3198, + "step": 36160 + }, + { + "epoch": 1.3965790184949225, + "grad_norm": 0.1592467725276947, + "learning_rate": 0.00010689730620229868, + "loss": 0.2591, + "step": 36170 + }, + { + "epoch": 1.3969651337889495, + "grad_norm": 0.4579029083251953, + "learning_rate": 0.0001068715651826969, + "loss": 0.2984, + "step": 36180 + }, + { + "epoch": 1.3973512490829763, + "grad_norm": 0.4286015033721924, + "learning_rate": 0.0001068458241630951, + "loss": 0.4098, + "step": 36190 + }, + { + "epoch": 1.397737364377003, + "grad_norm": 1.7824127674102783, + "learning_rate": 0.00010682008314349332, + "loss": 0.2446, + "step": 36200 + }, + { + "epoch": 1.3981234796710298, + "grad_norm": 0.8584449887275696, + "learning_rate": 0.00010679434212389152, + "loss": 0.1415, + "step": 36210 + }, + { + "epoch": 1.3985095949650566, + "grad_norm": 1.1699339151382446, + "learning_rate": 0.00010676860110428975, + "loss": 0.1706, + "step": 36220 + }, + { + "epoch": 1.3988957102590833, + "grad_norm": 2.615877389907837, + "learning_rate": 0.00010674286008468796, + "loss": 0.2605, + "step": 36230 + }, + { + "epoch": 1.39928182555311, + "grad_norm": 2.182037591934204, + "learning_rate": 0.00010671711906508617, + "loss": 0.2067, + "step": 36240 + }, + { + "epoch": 1.399667940847137, + "grad_norm": 2.183263063430786, + "learning_rate": 0.00010669137804548439, + "loss": 0.4704, + "step": 36250 + }, + { + "epoch": 1.4000540561411636, + "grad_norm": 3.505791187286377, + "learning_rate": 0.00010666563702588259, + "loss": 0.3385, + "step": 36260 + }, + { + "epoch": 1.4004401714351906, + "grad_norm": 1.2262030839920044, + "learning_rate": 0.00010663989600628081, + "loss": 0.3999, + "step": 36270 + }, + { + "epoch": 1.4008262867292174, + "grad_norm": 2.4024577140808105, + "learning_rate": 0.00010661415498667904, + "loss": 0.1984, + "step": 36280 + }, + { + "epoch": 1.4012124020232442, + "grad_norm": 0.4166090786457062, + "learning_rate": 0.00010658841396707724, + "loss": 0.2448, + "step": 36290 + }, + { + "epoch": 1.401598517317271, + "grad_norm": 0.422590047121048, + "learning_rate": 0.00010656267294747547, + "loss": 0.2464, + "step": 36300 + }, + { + "epoch": 1.4019846326112977, + "grad_norm": 2.287503480911255, + "learning_rate": 0.00010653693192787367, + "loss": 0.1621, + "step": 36310 + }, + { + "epoch": 1.4023707479053245, + "grad_norm": 0.8126110434532166, + "learning_rate": 0.00010651119090827188, + "loss": 0.2097, + "step": 36320 + }, + { + "epoch": 1.4027568631993512, + "grad_norm": 0.683016836643219, + "learning_rate": 0.00010648544988867008, + "loss": 0.1512, + "step": 36330 + }, + { + "epoch": 1.4031429784933782, + "grad_norm": 1.5477893352508545, + "learning_rate": 0.00010645970886906831, + "loss": 0.2169, + "step": 36340 + }, + { + "epoch": 1.403529093787405, + "grad_norm": 2.183166265487671, + "learning_rate": 0.00010643396784946653, + "loss": 0.4307, + "step": 36350 + }, + { + "epoch": 1.4039152090814317, + "grad_norm": 1.5782747268676758, + "learning_rate": 0.00010640822682986473, + "loss": 0.291, + "step": 36360 + }, + { + "epoch": 1.4043013243754585, + "grad_norm": 1.2823392152786255, + "learning_rate": 0.00010638248581026296, + "loss": 0.315, + "step": 36370 + }, + { + "epoch": 1.4046874396694853, + "grad_norm": 2.0126500129699707, + "learning_rate": 0.00010635674479066116, + "loss": 0.2565, + "step": 36380 + }, + { + "epoch": 1.405073554963512, + "grad_norm": 2.490217447280884, + "learning_rate": 0.00010633100377105937, + "loss": 0.2665, + "step": 36390 + }, + { + "epoch": 1.4054596702575388, + "grad_norm": 1.2054855823516846, + "learning_rate": 0.00010630526275145757, + "loss": 0.5455, + "step": 36400 + }, + { + "epoch": 1.4058457855515658, + "grad_norm": 0.7968757748603821, + "learning_rate": 0.0001062795217318558, + "loss": 0.333, + "step": 36410 + }, + { + "epoch": 1.4062319008455926, + "grad_norm": 1.1027718782424927, + "learning_rate": 0.00010625378071225403, + "loss": 0.2019, + "step": 36420 + }, + { + "epoch": 1.4066180161396193, + "grad_norm": 1.987302541732788, + "learning_rate": 0.00010622803969265223, + "loss": 0.3159, + "step": 36430 + }, + { + "epoch": 1.407004131433646, + "grad_norm": 1.5426512956619263, + "learning_rate": 0.00010620229867305045, + "loss": 0.3759, + "step": 36440 + }, + { + "epoch": 1.4073902467276729, + "grad_norm": 0.5228156447410583, + "learning_rate": 0.00010617655765344865, + "loss": 0.2465, + "step": 36450 + }, + { + "epoch": 1.4077763620216996, + "grad_norm": 0.46890121698379517, + "learning_rate": 0.00010615081663384687, + "loss": 0.302, + "step": 36460 + }, + { + "epoch": 1.4081624773157264, + "grad_norm": 2.1506495475769043, + "learning_rate": 0.0001061250756142451, + "loss": 0.2569, + "step": 36470 + }, + { + "epoch": 1.4085485926097534, + "grad_norm": 2.307468891143799, + "learning_rate": 0.0001060993345946433, + "loss": 0.2009, + "step": 36480 + }, + { + "epoch": 1.40893470790378, + "grad_norm": 0.07033026963472366, + "learning_rate": 0.00010607359357504152, + "loss": 0.1728, + "step": 36490 + }, + { + "epoch": 1.409320823197807, + "grad_norm": 0.3262972831726074, + "learning_rate": 0.00010604785255543972, + "loss": 0.2905, + "step": 36500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.222696300544e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e40ce92205791543b2b569ed2d085562424b146b --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c409e6e303ecd5508b7a55d1e95cd2566e1a09d6594b689cd7add25b66f0fd7a +size 291504128 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..962b4a0ebf67e9ef3359701a2bd4b01eca0881e2 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b36c1cc3f867592e6c79968a2d1c962a7de188656cfc1c4b2e6faa81dbfdaaa3 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..8392fb72aca9ce1fa59cd0b685ec87aa72b3542f Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..2968499a04095e0ee669cd8347b8cae8615bb392 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..bd4155eae3e28c32530342d6244cdcdcef746e73 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/trainer_state.json @@ -0,0 +1,25934 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.4286265878991466, + "eval_steps": 500, + "global_step": 37000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + }, + { + "epoch": 1.274566585582455, + "grad_norm": 0.5019121766090393, + "learning_rate": 0.0001150314683964632, + "loss": 0.1857, + "step": 33010 + }, + { + "epoch": 1.2749527008764816, + "grad_norm": 0.258952260017395, + "learning_rate": 0.0001150057273768614, + "loss": 0.11, + "step": 33020 + }, + { + "epoch": 1.2753388161705086, + "grad_norm": 0.8540348410606384, + "learning_rate": 0.00011497998635725962, + "loss": 0.1852, + "step": 33030 + }, + { + "epoch": 1.2757249314645354, + "grad_norm": 0.08689398318529129, + "learning_rate": 0.00011495424533765783, + "loss": 0.2225, + "step": 33040 + }, + { + "epoch": 1.2761110467585621, + "grad_norm": 0.42253822088241577, + "learning_rate": 0.00011492850431805605, + "loss": 0.0751, + "step": 33050 + }, + { + "epoch": 1.276497162052589, + "grad_norm": 1.2964017391204834, + "learning_rate": 0.00011490276329845426, + "loss": 0.2384, + "step": 33060 + }, + { + "epoch": 1.2768832773466157, + "grad_norm": 0.5337836146354675, + "learning_rate": 0.00011487702227885246, + "loss": 0.1415, + "step": 33070 + }, + { + "epoch": 1.2772693926406424, + "grad_norm": 2.7771682739257812, + "learning_rate": 0.00011485128125925069, + "loss": 0.378, + "step": 33080 + }, + { + "epoch": 1.2776555079346692, + "grad_norm": 1.5107232332229614, + "learning_rate": 0.00011482554023964889, + "loss": 0.2482, + "step": 33090 + }, + { + "epoch": 1.2780416232286962, + "grad_norm": 0.6885499358177185, + "learning_rate": 0.00011479979922004711, + "loss": 0.2122, + "step": 33100 + }, + { + "epoch": 1.278427738522723, + "grad_norm": 0.9016557335853577, + "learning_rate": 0.00011477405820044533, + "loss": 0.2841, + "step": 33110 + }, + { + "epoch": 1.2788138538167497, + "grad_norm": 1.9532525539398193, + "learning_rate": 0.00011474831718084354, + "loss": 0.2281, + "step": 33120 + }, + { + "epoch": 1.2791999691107765, + "grad_norm": 2.1078782081604004, + "learning_rate": 0.00011472257616124175, + "loss": 0.2836, + "step": 33130 + }, + { + "epoch": 1.2795860844048033, + "grad_norm": 0.19830390810966492, + "learning_rate": 0.00011469683514163995, + "loss": 0.388, + "step": 33140 + }, + { + "epoch": 1.27997219969883, + "grad_norm": 0.17538850009441376, + "learning_rate": 0.00011467109412203818, + "loss": 0.3274, + "step": 33150 + }, + { + "epoch": 1.2803583149928568, + "grad_norm": 0.7402139902114868, + "learning_rate": 0.00011464535310243641, + "loss": 0.1979, + "step": 33160 + }, + { + "epoch": 1.2807444302868838, + "grad_norm": 0.2097146362066269, + "learning_rate": 0.00011461961208283461, + "loss": 0.2464, + "step": 33170 + }, + { + "epoch": 1.2811305455809103, + "grad_norm": 1.2441083192825317, + "learning_rate": 0.00011459387106323283, + "loss": 0.266, + "step": 33180 + }, + { + "epoch": 1.2815166608749373, + "grad_norm": 2.518852710723877, + "learning_rate": 0.00011456813004363103, + "loss": 0.253, + "step": 33190 + }, + { + "epoch": 1.281902776168964, + "grad_norm": 0.8078998327255249, + "learning_rate": 0.00011454238902402925, + "loss": 0.2361, + "step": 33200 + }, + { + "epoch": 1.2822888914629909, + "grad_norm": 1.2297371625900269, + "learning_rate": 0.00011451664800442745, + "loss": 0.1974, + "step": 33210 + }, + { + "epoch": 1.2826750067570176, + "grad_norm": 0.4303855895996094, + "learning_rate": 0.00011449090698482567, + "loss": 0.3563, + "step": 33220 + }, + { + "epoch": 1.2830611220510444, + "grad_norm": 1.3215210437774658, + "learning_rate": 0.0001144651659652239, + "loss": 0.2818, + "step": 33230 + }, + { + "epoch": 1.2834472373450712, + "grad_norm": 1.546265959739685, + "learning_rate": 0.0001144394249456221, + "loss": 0.5778, + "step": 33240 + }, + { + "epoch": 1.283833352639098, + "grad_norm": 0.8895953297615051, + "learning_rate": 0.00011441368392602033, + "loss": 0.2485, + "step": 33250 + }, + { + "epoch": 1.284219467933125, + "grad_norm": 0.7534870505332947, + "learning_rate": 0.00011438794290641853, + "loss": 0.2825, + "step": 33260 + }, + { + "epoch": 1.2846055832271517, + "grad_norm": 0.052820973098278046, + "learning_rate": 0.00011436220188681674, + "loss": 0.2191, + "step": 33270 + }, + { + "epoch": 1.2849916985211784, + "grad_norm": 0.9264475107192993, + "learning_rate": 0.00011433646086721494, + "loss": 0.181, + "step": 33280 + }, + { + "epoch": 1.2853778138152052, + "grad_norm": 0.2128441333770752, + "learning_rate": 0.00011431071984761317, + "loss": 0.1819, + "step": 33290 + }, + { + "epoch": 1.285763929109232, + "grad_norm": 0.5400950312614441, + "learning_rate": 0.0001142849788280114, + "loss": 0.4316, + "step": 33300 + }, + { + "epoch": 1.2861500444032588, + "grad_norm": 0.8033271431922913, + "learning_rate": 0.00011425923780840959, + "loss": 0.2146, + "step": 33310 + }, + { + "epoch": 1.2865361596972855, + "grad_norm": 2.012575149536133, + "learning_rate": 0.00011423349678880782, + "loss": 0.4335, + "step": 33320 + }, + { + "epoch": 1.2869222749913125, + "grad_norm": 0.7352376580238342, + "learning_rate": 0.00011420775576920602, + "loss": 0.2124, + "step": 33330 + }, + { + "epoch": 1.2873083902853393, + "grad_norm": 0.769036591053009, + "learning_rate": 0.00011418201474960423, + "loss": 0.3602, + "step": 33340 + }, + { + "epoch": 1.287694505579366, + "grad_norm": 0.250592976808548, + "learning_rate": 0.00011415627373000246, + "loss": 0.1692, + "step": 33350 + }, + { + "epoch": 1.2880806208733928, + "grad_norm": 2.43820858001709, + "learning_rate": 0.00011413053271040066, + "loss": 0.2777, + "step": 33360 + }, + { + "epoch": 1.2884667361674196, + "grad_norm": 1.3179954290390015, + "learning_rate": 0.00011410479169079889, + "loss": 0.1794, + "step": 33370 + }, + { + "epoch": 1.2888528514614463, + "grad_norm": 1.0040466785430908, + "learning_rate": 0.00011407905067119709, + "loss": 0.3037, + "step": 33380 + }, + { + "epoch": 1.2892389667554731, + "grad_norm": 5.296288013458252, + "learning_rate": 0.00011405330965159531, + "loss": 0.2904, + "step": 33390 + }, + { + "epoch": 1.2896250820495, + "grad_norm": 0.4267273247241974, + "learning_rate": 0.00011402756863199351, + "loss": 0.2263, + "step": 33400 + }, + { + "epoch": 1.2900111973435266, + "grad_norm": 0.8817713260650635, + "learning_rate": 0.00011400182761239173, + "loss": 0.2715, + "step": 33410 + }, + { + "epoch": 1.2903973126375536, + "grad_norm": 2.7891275882720947, + "learning_rate": 0.00011397608659278995, + "loss": 0.2781, + "step": 33420 + }, + { + "epoch": 1.2907834279315804, + "grad_norm": 0.3013952374458313, + "learning_rate": 0.00011395034557318815, + "loss": 0.2563, + "step": 33430 + }, + { + "epoch": 1.2911695432256072, + "grad_norm": 1.766413927078247, + "learning_rate": 0.00011392460455358638, + "loss": 0.1854, + "step": 33440 + }, + { + "epoch": 1.291555658519634, + "grad_norm": 0.25331103801727295, + "learning_rate": 0.00011389886353398458, + "loss": 0.1206, + "step": 33450 + }, + { + "epoch": 1.2919417738136607, + "grad_norm": 0.036400288343429565, + "learning_rate": 0.0001138731225143828, + "loss": 0.4707, + "step": 33460 + }, + { + "epoch": 1.2923278891076875, + "grad_norm": 1.5620888471603394, + "learning_rate": 0.00011384738149478102, + "loss": 0.3312, + "step": 33470 + }, + { + "epoch": 1.2927140044017142, + "grad_norm": 0.6670392155647278, + "learning_rate": 0.00011382164047517922, + "loss": 0.2341, + "step": 33480 + }, + { + "epoch": 1.2931001196957412, + "grad_norm": 2.3108737468719482, + "learning_rate": 0.00011379589945557745, + "loss": 0.3843, + "step": 33490 + }, + { + "epoch": 1.293486234989768, + "grad_norm": 0.8025147318840027, + "learning_rate": 0.00011377015843597565, + "loss": 0.1982, + "step": 33500 + }, + { + "epoch": 1.2938723502837948, + "grad_norm": 1.7835719585418701, + "learning_rate": 0.00011374441741637387, + "loss": 0.3285, + "step": 33510 + }, + { + "epoch": 1.2942584655778215, + "grad_norm": 2.041508913040161, + "learning_rate": 0.00011371867639677207, + "loss": 0.2044, + "step": 33520 + }, + { + "epoch": 1.2946445808718483, + "grad_norm": 1.103378415107727, + "learning_rate": 0.0001136929353771703, + "loss": 0.1682, + "step": 33530 + }, + { + "epoch": 1.295030696165875, + "grad_norm": 0.057376351207494736, + "learning_rate": 0.00011366719435756851, + "loss": 0.1642, + "step": 33540 + }, + { + "epoch": 1.2954168114599018, + "grad_norm": 0.6539410948753357, + "learning_rate": 0.00011364145333796671, + "loss": 0.1549, + "step": 33550 + }, + { + "epoch": 1.2958029267539288, + "grad_norm": 1.250543236732483, + "learning_rate": 0.00011361571231836494, + "loss": 0.3764, + "step": 33560 + }, + { + "epoch": 1.2961890420479556, + "grad_norm": 0.23697887361049652, + "learning_rate": 0.00011358997129876314, + "loss": 0.3999, + "step": 33570 + }, + { + "epoch": 1.2965751573419824, + "grad_norm": 0.9318505525588989, + "learning_rate": 0.00011356423027916137, + "loss": 0.4156, + "step": 33580 + }, + { + "epoch": 1.2969612726360091, + "grad_norm": 1.3910777568817139, + "learning_rate": 0.00011353848925955957, + "loss": 0.3455, + "step": 33590 + }, + { + "epoch": 1.297347387930036, + "grad_norm": 1.6764451265335083, + "learning_rate": 0.00011351274823995779, + "loss": 0.1884, + "step": 33600 + }, + { + "epoch": 1.2977335032240627, + "grad_norm": 0.9300051927566528, + "learning_rate": 0.000113487007220356, + "loss": 0.119, + "step": 33610 + }, + { + "epoch": 1.2981196185180894, + "grad_norm": 2.447462558746338, + "learning_rate": 0.00011346126620075422, + "loss": 0.4403, + "step": 33620 + }, + { + "epoch": 1.2985057338121164, + "grad_norm": 1.216407060623169, + "learning_rate": 0.00011343552518115243, + "loss": 0.2415, + "step": 33630 + }, + { + "epoch": 1.298891849106143, + "grad_norm": 2.968648910522461, + "learning_rate": 0.00011340978416155063, + "loss": 0.2899, + "step": 33640 + }, + { + "epoch": 1.29927796440017, + "grad_norm": 0.6649970412254333, + "learning_rate": 0.00011338404314194886, + "loss": 0.3809, + "step": 33650 + }, + { + "epoch": 1.2996640796941967, + "grad_norm": 1.7277917861938477, + "learning_rate": 0.00011335830212234709, + "loss": 0.3308, + "step": 33660 + }, + { + "epoch": 1.3000501949882235, + "grad_norm": 1.3269709348678589, + "learning_rate": 0.00011333256110274529, + "loss": 0.3682, + "step": 33670 + }, + { + "epoch": 1.3004363102822503, + "grad_norm": 0.20609407126903534, + "learning_rate": 0.0001133068200831435, + "loss": 0.1379, + "step": 33680 + }, + { + "epoch": 1.300822425576277, + "grad_norm": 0.6592215299606323, + "learning_rate": 0.00011328107906354171, + "loss": 0.2746, + "step": 33690 + }, + { + "epoch": 1.3012085408703038, + "grad_norm": 1.903635859489441, + "learning_rate": 0.00011325533804393993, + "loss": 0.4729, + "step": 33700 + }, + { + "epoch": 1.3015946561643306, + "grad_norm": 0.8432504534721375, + "learning_rate": 0.00011322959702433813, + "loss": 0.2835, + "step": 33710 + }, + { + "epoch": 1.3019807714583576, + "grad_norm": 0.9862542152404785, + "learning_rate": 0.00011320385600473635, + "loss": 0.1823, + "step": 33720 + }, + { + "epoch": 1.3023668867523843, + "grad_norm": 3.845738649368286, + "learning_rate": 0.00011317811498513458, + "loss": 0.2542, + "step": 33730 + }, + { + "epoch": 1.302753002046411, + "grad_norm": 0.6317747235298157, + "learning_rate": 0.00011315237396553278, + "loss": 0.22, + "step": 33740 + }, + { + "epoch": 1.3031391173404379, + "grad_norm": 2.5221354961395264, + "learning_rate": 0.000113126632945931, + "loss": 0.2253, + "step": 33750 + }, + { + "epoch": 1.3035252326344646, + "grad_norm": 1.3326247930526733, + "learning_rate": 0.0001131008919263292, + "loss": 0.2021, + "step": 33760 + }, + { + "epoch": 1.3039113479284914, + "grad_norm": 1.148047685623169, + "learning_rate": 0.00011307515090672742, + "loss": 0.3987, + "step": 33770 + }, + { + "epoch": 1.3042974632225182, + "grad_norm": 0.19721268117427826, + "learning_rate": 0.00011304940988712562, + "loss": 0.2642, + "step": 33780 + }, + { + "epoch": 1.3046835785165452, + "grad_norm": 1.4060617685317993, + "learning_rate": 0.00011302366886752385, + "loss": 0.2736, + "step": 33790 + }, + { + "epoch": 1.305069693810572, + "grad_norm": 1.0736548900604248, + "learning_rate": 0.00011299792784792207, + "loss": 0.2448, + "step": 33800 + }, + { + "epoch": 1.3054558091045987, + "grad_norm": 4.352476119995117, + "learning_rate": 0.00011297218682832027, + "loss": 0.383, + "step": 33810 + }, + { + "epoch": 1.3058419243986255, + "grad_norm": 0.2249228060245514, + "learning_rate": 0.0001129464458087185, + "loss": 0.14, + "step": 33820 + }, + { + "epoch": 1.3062280396926522, + "grad_norm": 0.4820781946182251, + "learning_rate": 0.0001129207047891167, + "loss": 0.248, + "step": 33830 + }, + { + "epoch": 1.306614154986679, + "grad_norm": 2.2983391284942627, + "learning_rate": 0.00011289496376951491, + "loss": 0.2608, + "step": 33840 + }, + { + "epoch": 1.3070002702807058, + "grad_norm": 1.3315671682357788, + "learning_rate": 0.00011286922274991314, + "loss": 0.1727, + "step": 33850 + }, + { + "epoch": 1.3073863855747327, + "grad_norm": 2.060299873352051, + "learning_rate": 0.00011284348173031134, + "loss": 0.3097, + "step": 33860 + }, + { + "epoch": 1.3077725008687593, + "grad_norm": 2.096285581588745, + "learning_rate": 0.00011281774071070957, + "loss": 0.2305, + "step": 33870 + }, + { + "epoch": 1.3081586161627863, + "grad_norm": 0.4997636675834656, + "learning_rate": 0.00011279199969110777, + "loss": 0.1993, + "step": 33880 + }, + { + "epoch": 1.308544731456813, + "grad_norm": 0.58636474609375, + "learning_rate": 0.00011276625867150599, + "loss": 0.1873, + "step": 33890 + }, + { + "epoch": 1.3089308467508398, + "grad_norm": 0.9128592610359192, + "learning_rate": 0.00011274051765190419, + "loss": 0.1885, + "step": 33900 + }, + { + "epoch": 1.3093169620448666, + "grad_norm": 2.228043794631958, + "learning_rate": 0.0001127147766323024, + "loss": 0.3649, + "step": 33910 + }, + { + "epoch": 1.3097030773388934, + "grad_norm": 1.069002389907837, + "learning_rate": 0.00011268903561270063, + "loss": 0.5454, + "step": 33920 + }, + { + "epoch": 1.3100891926329201, + "grad_norm": 0.6207597851753235, + "learning_rate": 0.00011266329459309883, + "loss": 0.2329, + "step": 33930 + }, + { + "epoch": 1.3104753079269469, + "grad_norm": 1.262247920036316, + "learning_rate": 0.00011263755357349706, + "loss": 0.3437, + "step": 33940 + }, + { + "epoch": 1.3108614232209739, + "grad_norm": 1.7429994344711304, + "learning_rate": 0.00011261181255389526, + "loss": 0.228, + "step": 33950 + }, + { + "epoch": 1.3112475385150006, + "grad_norm": 0.646900475025177, + "learning_rate": 0.00011258607153429349, + "loss": 0.3739, + "step": 33960 + }, + { + "epoch": 1.3116336538090274, + "grad_norm": 1.8228782415390015, + "learning_rate": 0.0001125603305146917, + "loss": 0.2325, + "step": 33970 + }, + { + "epoch": 1.3120197691030542, + "grad_norm": 3.539228916168213, + "learning_rate": 0.0001125345894950899, + "loss": 0.194, + "step": 33980 + }, + { + "epoch": 1.312405884397081, + "grad_norm": 1.2801135778427124, + "learning_rate": 0.00011250884847548813, + "loss": 0.3069, + "step": 33990 + }, + { + "epoch": 1.3127919996911077, + "grad_norm": 3.6265695095062256, + "learning_rate": 0.00011248310745588633, + "loss": 0.3113, + "step": 34000 + }, + { + "epoch": 1.3131781149851345, + "grad_norm": 0.07370063662528992, + "learning_rate": 0.00011245736643628455, + "loss": 0.1449, + "step": 34010 + }, + { + "epoch": 1.3135642302791615, + "grad_norm": 1.0295637845993042, + "learning_rate": 0.00011243162541668275, + "loss": 0.22, + "step": 34020 + }, + { + "epoch": 1.313950345573188, + "grad_norm": 0.8803662061691284, + "learning_rate": 0.00011240588439708098, + "loss": 0.1368, + "step": 34030 + }, + { + "epoch": 1.314336460867215, + "grad_norm": 1.6597707271575928, + "learning_rate": 0.00011238014337747919, + "loss": 0.3038, + "step": 34040 + }, + { + "epoch": 1.3147225761612418, + "grad_norm": 2.115492343902588, + "learning_rate": 0.00011235440235787739, + "loss": 0.1754, + "step": 34050 + }, + { + "epoch": 1.3151086914552685, + "grad_norm": 0.8143919706344604, + "learning_rate": 0.00011232866133827562, + "loss": 0.3764, + "step": 34060 + }, + { + "epoch": 1.3154948067492953, + "grad_norm": 0.14369767904281616, + "learning_rate": 0.00011230292031867382, + "loss": 0.1742, + "step": 34070 + }, + { + "epoch": 1.315880922043322, + "grad_norm": 1.0129845142364502, + "learning_rate": 0.00011227717929907205, + "loss": 0.1458, + "step": 34080 + }, + { + "epoch": 1.316267037337349, + "grad_norm": 2.7300291061401367, + "learning_rate": 0.00011225143827947025, + "loss": 0.3939, + "step": 34090 + }, + { + "epoch": 1.3166531526313756, + "grad_norm": 0.20205609500408173, + "learning_rate": 0.00011222569725986847, + "loss": 0.29, + "step": 34100 + }, + { + "epoch": 1.3170392679254026, + "grad_norm": 1.8928464651107788, + "learning_rate": 0.00011219995624026669, + "loss": 0.1742, + "step": 34110 + }, + { + "epoch": 1.3174253832194294, + "grad_norm": 0.2639687955379486, + "learning_rate": 0.00011217421522066488, + "loss": 0.1745, + "step": 34120 + }, + { + "epoch": 1.3178114985134561, + "grad_norm": 0.5906389355659485, + "learning_rate": 0.00011214847420106311, + "loss": 0.2134, + "step": 34130 + }, + { + "epoch": 1.318197613807483, + "grad_norm": 0.9190629720687866, + "learning_rate": 0.00011212273318146131, + "loss": 0.2547, + "step": 34140 + }, + { + "epoch": 1.3185837291015097, + "grad_norm": 0.5749151110649109, + "learning_rate": 0.00011209699216185954, + "loss": 0.1688, + "step": 34150 + }, + { + "epoch": 1.3189698443955364, + "grad_norm": 0.82295823097229, + "learning_rate": 0.00011207125114225777, + "loss": 0.2884, + "step": 34160 + }, + { + "epoch": 1.3193559596895632, + "grad_norm": 0.07816460728645325, + "learning_rate": 0.00011204551012265597, + "loss": 0.2418, + "step": 34170 + }, + { + "epoch": 1.3197420749835902, + "grad_norm": 0.6417407393455505, + "learning_rate": 0.00011201976910305418, + "loss": 0.2557, + "step": 34180 + }, + { + "epoch": 1.320128190277617, + "grad_norm": 6.093267440795898, + "learning_rate": 0.00011199402808345238, + "loss": 0.3088, + "step": 34190 + }, + { + "epoch": 1.3205143055716437, + "grad_norm": 1.8861887454986572, + "learning_rate": 0.0001119682870638506, + "loss": 0.2204, + "step": 34200 + }, + { + "epoch": 1.3209004208656705, + "grad_norm": 2.3272714614868164, + "learning_rate": 0.0001119425460442488, + "loss": 0.2236, + "step": 34210 + }, + { + "epoch": 1.3212865361596973, + "grad_norm": 0.9608810544013977, + "learning_rate": 0.00011191680502464703, + "loss": 0.1897, + "step": 34220 + }, + { + "epoch": 1.321672651453724, + "grad_norm": 1.2157350778579712, + "learning_rate": 0.00011189106400504526, + "loss": 0.1526, + "step": 34230 + }, + { + "epoch": 1.3220587667477508, + "grad_norm": 1.6684671640396118, + "learning_rate": 0.00011186532298544346, + "loss": 0.3394, + "step": 34240 + }, + { + "epoch": 1.3224448820417778, + "grad_norm": 2.0432374477386475, + "learning_rate": 0.00011183958196584167, + "loss": 0.2183, + "step": 34250 + }, + { + "epoch": 1.3228309973358043, + "grad_norm": 0.9436892867088318, + "learning_rate": 0.00011181384094623988, + "loss": 0.2947, + "step": 34260 + }, + { + "epoch": 1.3232171126298313, + "grad_norm": 0.23260092735290527, + "learning_rate": 0.0001117880999266381, + "loss": 0.114, + "step": 34270 + }, + { + "epoch": 1.323603227923858, + "grad_norm": 1.2291594743728638, + "learning_rate": 0.0001117623589070363, + "loss": 0.3145, + "step": 34280 + }, + { + "epoch": 1.3239893432178849, + "grad_norm": 0.41411107778549194, + "learning_rate": 0.00011173661788743452, + "loss": 0.2937, + "step": 34290 + }, + { + "epoch": 1.3243754585119116, + "grad_norm": 2.354405164718628, + "learning_rate": 0.00011171087686783275, + "loss": 0.3933, + "step": 34300 + }, + { + "epoch": 1.3247615738059384, + "grad_norm": 2.6997978687286377, + "learning_rate": 0.00011168513584823095, + "loss": 0.1494, + "step": 34310 + }, + { + "epoch": 1.3251476890999654, + "grad_norm": 2.8430919647216797, + "learning_rate": 0.00011165939482862916, + "loss": 0.2869, + "step": 34320 + }, + { + "epoch": 1.325533804393992, + "grad_norm": 1.1737356185913086, + "learning_rate": 0.00011163365380902738, + "loss": 0.2792, + "step": 34330 + }, + { + "epoch": 1.325919919688019, + "grad_norm": 4.123973846435547, + "learning_rate": 0.00011160791278942559, + "loss": 0.5211, + "step": 34340 + }, + { + "epoch": 1.3263060349820457, + "grad_norm": 0.8862038850784302, + "learning_rate": 0.00011158217176982382, + "loss": 0.2976, + "step": 34350 + }, + { + "epoch": 1.3266921502760725, + "grad_norm": 1.8690590858459473, + "learning_rate": 0.00011155643075022202, + "loss": 0.2485, + "step": 34360 + }, + { + "epoch": 1.3270782655700992, + "grad_norm": 2.885589599609375, + "learning_rate": 0.00011153068973062024, + "loss": 0.3084, + "step": 34370 + }, + { + "epoch": 1.327464380864126, + "grad_norm": 0.9898788928985596, + "learning_rate": 0.00011150494871101844, + "loss": 0.261, + "step": 34380 + }, + { + "epoch": 1.3278504961581528, + "grad_norm": 0.6879653930664062, + "learning_rate": 0.00011147920769141667, + "loss": 0.2082, + "step": 34390 + }, + { + "epoch": 1.3282366114521795, + "grad_norm": 1.2619003057479858, + "learning_rate": 0.00011145346667181487, + "loss": 0.2402, + "step": 34400 + }, + { + "epoch": 1.3286227267462065, + "grad_norm": 1.1212007999420166, + "learning_rate": 0.00011142772565221308, + "loss": 0.3525, + "step": 34410 + }, + { + "epoch": 1.3290088420402333, + "grad_norm": 1.8431956768035889, + "learning_rate": 0.00011140198463261131, + "loss": 0.2212, + "step": 34420 + }, + { + "epoch": 1.32939495733426, + "grad_norm": 0.6185423731803894, + "learning_rate": 0.00011137624361300951, + "loss": 0.2455, + "step": 34430 + }, + { + "epoch": 1.3297810726282868, + "grad_norm": 2.3791301250457764, + "learning_rate": 0.00011135050259340774, + "loss": 0.1763, + "step": 34440 + }, + { + "epoch": 1.3301671879223136, + "grad_norm": 0.4928603768348694, + "learning_rate": 0.00011132476157380594, + "loss": 0.2381, + "step": 34450 + }, + { + "epoch": 1.3305533032163404, + "grad_norm": 1.5636029243469238, + "learning_rate": 0.00011129902055420416, + "loss": 0.1368, + "step": 34460 + }, + { + "epoch": 1.3309394185103671, + "grad_norm": 0.9425283074378967, + "learning_rate": 0.00011127327953460238, + "loss": 0.2825, + "step": 34470 + }, + { + "epoch": 1.3313255338043941, + "grad_norm": 1.2257115840911865, + "learning_rate": 0.00011124753851500058, + "loss": 0.2547, + "step": 34480 + }, + { + "epoch": 1.3317116490984207, + "grad_norm": 0.9416170716285706, + "learning_rate": 0.0001112217974953988, + "loss": 0.2766, + "step": 34490 + }, + { + "epoch": 1.3320977643924476, + "grad_norm": 0.5123847126960754, + "learning_rate": 0.000111196056475797, + "loss": 0.4733, + "step": 34500 + }, + { + "epoch": 1.3324838796864744, + "grad_norm": 1.5581384897232056, + "learning_rate": 0.00011117031545619523, + "loss": 0.1597, + "step": 34510 + }, + { + "epoch": 1.3328699949805012, + "grad_norm": 2.377333879470825, + "learning_rate": 0.00011114457443659343, + "loss": 0.209, + "step": 34520 + }, + { + "epoch": 1.333256110274528, + "grad_norm": 1.7840913534164429, + "learning_rate": 0.00011111883341699166, + "loss": 0.1759, + "step": 34530 + }, + { + "epoch": 1.3336422255685547, + "grad_norm": 1.1825993061065674, + "learning_rate": 0.00011109309239738987, + "loss": 0.2464, + "step": 34540 + }, + { + "epoch": 1.3340283408625815, + "grad_norm": 1.8859659433364868, + "learning_rate": 0.00011106735137778807, + "loss": 0.3539, + "step": 34550 + }, + { + "epoch": 1.3344144561566083, + "grad_norm": 1.9698175191879272, + "learning_rate": 0.0001110416103581863, + "loss": 0.3301, + "step": 34560 + }, + { + "epoch": 1.3348005714506352, + "grad_norm": 0.7649385333061218, + "learning_rate": 0.0001110158693385845, + "loss": 0.232, + "step": 34570 + }, + { + "epoch": 1.335186686744662, + "grad_norm": 0.56386399269104, + "learning_rate": 0.00011099012831898272, + "loss": 0.3425, + "step": 34580 + }, + { + "epoch": 1.3355728020386888, + "grad_norm": 2.956003189086914, + "learning_rate": 0.00011096438729938092, + "loss": 0.1518, + "step": 34590 + }, + { + "epoch": 1.3359589173327155, + "grad_norm": 2.612029552459717, + "learning_rate": 0.00011093864627977915, + "loss": 0.2765, + "step": 34600 + }, + { + "epoch": 1.3363450326267423, + "grad_norm": 0.9674397706985474, + "learning_rate": 0.00011091290526017736, + "loss": 0.303, + "step": 34610 + }, + { + "epoch": 1.336731147920769, + "grad_norm": 0.9578921794891357, + "learning_rate": 0.00011088716424057556, + "loss": 0.1405, + "step": 34620 + }, + { + "epoch": 1.3371172632147958, + "grad_norm": 2.168065071105957, + "learning_rate": 0.00011086142322097379, + "loss": 0.1914, + "step": 34630 + }, + { + "epoch": 1.3375033785088228, + "grad_norm": 1.3166526556015015, + "learning_rate": 0.00011083568220137199, + "loss": 0.4134, + "step": 34640 + }, + { + "epoch": 1.3378894938028496, + "grad_norm": 0.9082283973693848, + "learning_rate": 0.00011080994118177022, + "loss": 0.2693, + "step": 34650 + }, + { + "epoch": 1.3382756090968764, + "grad_norm": 2.203007698059082, + "learning_rate": 0.00011078420016216844, + "loss": 0.1847, + "step": 34660 + }, + { + "epoch": 1.3386617243909031, + "grad_norm": 0.8101674914360046, + "learning_rate": 0.00011075845914256664, + "loss": 0.3111, + "step": 34670 + }, + { + "epoch": 1.33904783968493, + "grad_norm": 1.9545695781707764, + "learning_rate": 0.00011073271812296486, + "loss": 0.3761, + "step": 34680 + }, + { + "epoch": 1.3394339549789567, + "grad_norm": 1.547581672668457, + "learning_rate": 0.00011070697710336306, + "loss": 0.2374, + "step": 34690 + }, + { + "epoch": 1.3398200702729834, + "grad_norm": 3.3519034385681152, + "learning_rate": 0.00011068123608376128, + "loss": 0.1957, + "step": 34700 + }, + { + "epoch": 1.3402061855670104, + "grad_norm": 1.5508599281311035, + "learning_rate": 0.00011065549506415948, + "loss": 0.4171, + "step": 34710 + }, + { + "epoch": 1.340592300861037, + "grad_norm": 1.8547546863555908, + "learning_rate": 0.00011062975404455771, + "loss": 0.1872, + "step": 34720 + }, + { + "epoch": 1.340978416155064, + "grad_norm": 1.4600756168365479, + "learning_rate": 0.00011060401302495594, + "loss": 0.3515, + "step": 34730 + }, + { + "epoch": 1.3413645314490907, + "grad_norm": 0.05774044618010521, + "learning_rate": 0.00011057827200535414, + "loss": 0.1604, + "step": 34740 + }, + { + "epoch": 1.3417506467431175, + "grad_norm": 2.8793342113494873, + "learning_rate": 0.00011055253098575235, + "loss": 0.3095, + "step": 34750 + }, + { + "epoch": 1.3421367620371443, + "grad_norm": 2.241042375564575, + "learning_rate": 0.00011052678996615055, + "loss": 0.2511, + "step": 34760 + }, + { + "epoch": 1.342522877331171, + "grad_norm": 1.9320632219314575, + "learning_rate": 0.00011050104894654878, + "loss": 0.4493, + "step": 34770 + }, + { + "epoch": 1.3429089926251978, + "grad_norm": 1.6483882665634155, + "learning_rate": 0.000110475307926947, + "loss": 0.217, + "step": 34780 + }, + { + "epoch": 1.3432951079192246, + "grad_norm": 0.9635765552520752, + "learning_rate": 0.0001104495669073452, + "loss": 0.5458, + "step": 34790 + }, + { + "epoch": 1.3436812232132516, + "grad_norm": 1.2436567544937134, + "learning_rate": 0.00011042382588774343, + "loss": 0.2857, + "step": 34800 + }, + { + "epoch": 1.3440673385072783, + "grad_norm": 2.8082425594329834, + "learning_rate": 0.00011039808486814163, + "loss": 0.3439, + "step": 34810 + }, + { + "epoch": 1.344453453801305, + "grad_norm": 1.0430901050567627, + "learning_rate": 0.00011037234384853984, + "loss": 0.1404, + "step": 34820 + }, + { + "epoch": 1.3448395690953319, + "grad_norm": 1.7387149333953857, + "learning_rate": 0.00011034660282893806, + "loss": 0.395, + "step": 34830 + }, + { + "epoch": 1.3452256843893586, + "grad_norm": 1.2713748216629028, + "learning_rate": 0.00011032086180933627, + "loss": 0.29, + "step": 34840 + }, + { + "epoch": 1.3456117996833854, + "grad_norm": 0.26068204641342163, + "learning_rate": 0.0001102951207897345, + "loss": 0.1814, + "step": 34850 + }, + { + "epoch": 1.3459979149774122, + "grad_norm": 2.4163243770599365, + "learning_rate": 0.0001102693797701327, + "loss": 0.2445, + "step": 34860 + }, + { + "epoch": 1.3463840302714392, + "grad_norm": 2.2439687252044678, + "learning_rate": 0.00011024363875053092, + "loss": 0.3338, + "step": 34870 + }, + { + "epoch": 1.346770145565466, + "grad_norm": 0.2822403609752655, + "learning_rate": 0.00011021789773092912, + "loss": 0.1648, + "step": 34880 + }, + { + "epoch": 1.3471562608594927, + "grad_norm": 0.07319017499685287, + "learning_rate": 0.00011019215671132734, + "loss": 0.107, + "step": 34890 + }, + { + "epoch": 1.3475423761535195, + "grad_norm": 0.9809044003486633, + "learning_rate": 0.00011016641569172555, + "loss": 0.256, + "step": 34900 + }, + { + "epoch": 1.3479284914475462, + "grad_norm": 0.5016226768493652, + "learning_rate": 0.00011014067467212376, + "loss": 0.3229, + "step": 34910 + }, + { + "epoch": 1.348314606741573, + "grad_norm": 1.3026005029678345, + "learning_rate": 0.00011011493365252199, + "loss": 0.2145, + "step": 34920 + }, + { + "epoch": 1.3487007220355998, + "grad_norm": 1.0752215385437012, + "learning_rate": 0.00011008919263292019, + "loss": 0.2355, + "step": 34930 + }, + { + "epoch": 1.3490868373296268, + "grad_norm": 2.2703003883361816, + "learning_rate": 0.00011006345161331842, + "loss": 0.2079, + "step": 34940 + }, + { + "epoch": 1.3494729526236533, + "grad_norm": 1.1323810815811157, + "learning_rate": 0.00011003771059371662, + "loss": 0.1015, + "step": 34950 + }, + { + "epoch": 1.3498590679176803, + "grad_norm": 0.10813555121421814, + "learning_rate": 0.00011001196957411484, + "loss": 0.4214, + "step": 34960 + }, + { + "epoch": 1.350245183211707, + "grad_norm": 0.07815568149089813, + "learning_rate": 0.00010998622855451306, + "loss": 0.1211, + "step": 34970 + }, + { + "epoch": 1.3506312985057338, + "grad_norm": 0.6748234629631042, + "learning_rate": 0.00010996048753491126, + "loss": 0.3508, + "step": 34980 + }, + { + "epoch": 1.3510174137997606, + "grad_norm": 1.8556997776031494, + "learning_rate": 0.00010993474651530948, + "loss": 0.2268, + "step": 34990 + }, + { + "epoch": 1.3514035290937874, + "grad_norm": 0.8696061372756958, + "learning_rate": 0.00010990900549570768, + "loss": 0.4321, + "step": 35000 + }, + { + "epoch": 1.3517896443878141, + "grad_norm": 0.42442765831947327, + "learning_rate": 0.00010988326447610591, + "loss": 0.1944, + "step": 35010 + }, + { + "epoch": 1.352175759681841, + "grad_norm": 1.0474554300308228, + "learning_rate": 0.00010985752345650411, + "loss": 0.1342, + "step": 35020 + }, + { + "epoch": 1.3525618749758679, + "grad_norm": 0.607037365436554, + "learning_rate": 0.00010983178243690234, + "loss": 0.2965, + "step": 35030 + }, + { + "epoch": 1.3529479902698947, + "grad_norm": 1.8160990476608276, + "learning_rate": 0.00010980604141730055, + "loss": 0.3192, + "step": 35040 + }, + { + "epoch": 1.3533341055639214, + "grad_norm": 2.0026509761810303, + "learning_rate": 0.00010978030039769875, + "loss": 0.3054, + "step": 35050 + }, + { + "epoch": 1.3537202208579482, + "grad_norm": 0.9203600883483887, + "learning_rate": 0.00010975455937809698, + "loss": 0.253, + "step": 35060 + }, + { + "epoch": 1.354106336151975, + "grad_norm": 0.33198195695877075, + "learning_rate": 0.00010972881835849518, + "loss": 0.3885, + "step": 35070 + }, + { + "epoch": 1.3544924514460017, + "grad_norm": 0.3201223611831665, + "learning_rate": 0.0001097030773388934, + "loss": 0.3029, + "step": 35080 + }, + { + "epoch": 1.3548785667400285, + "grad_norm": 1.2589943408966064, + "learning_rate": 0.0001096773363192916, + "loss": 0.4243, + "step": 35090 + }, + { + "epoch": 1.3552646820340555, + "grad_norm": 1.5106219053268433, + "learning_rate": 0.00010965159529968983, + "loss": 0.2585, + "step": 35100 + }, + { + "epoch": 1.3556507973280822, + "grad_norm": 1.429799199104309, + "learning_rate": 0.00010962585428008804, + "loss": 0.1961, + "step": 35110 + }, + { + "epoch": 1.356036912622109, + "grad_norm": 2.1211297512054443, + "learning_rate": 0.00010960011326048624, + "loss": 0.4057, + "step": 35120 + }, + { + "epoch": 1.3564230279161358, + "grad_norm": 2.5154731273651123, + "learning_rate": 0.00010957437224088447, + "loss": 0.3787, + "step": 35130 + }, + { + "epoch": 1.3568091432101625, + "grad_norm": 0.4914834201335907, + "learning_rate": 0.00010954863122128267, + "loss": 0.234, + "step": 35140 + }, + { + "epoch": 1.3571952585041893, + "grad_norm": 0.26685893535614014, + "learning_rate": 0.0001095228902016809, + "loss": 0.2841, + "step": 35150 + }, + { + "epoch": 1.357581373798216, + "grad_norm": 0.15462155640125275, + "learning_rate": 0.00010949714918207912, + "loss": 0.2269, + "step": 35160 + }, + { + "epoch": 1.357967489092243, + "grad_norm": 1.3887063264846802, + "learning_rate": 0.00010947140816247732, + "loss": 0.3455, + "step": 35170 + }, + { + "epoch": 1.3583536043862696, + "grad_norm": 0.786374032497406, + "learning_rate": 0.00010944566714287554, + "loss": 0.2897, + "step": 35180 + }, + { + "epoch": 1.3587397196802966, + "grad_norm": 1.100475549697876, + "learning_rate": 0.00010941992612327374, + "loss": 0.2892, + "step": 35190 + }, + { + "epoch": 1.3591258349743234, + "grad_norm": 0.7676102519035339, + "learning_rate": 0.00010939418510367196, + "loss": 0.1942, + "step": 35200 + }, + { + "epoch": 1.3595119502683501, + "grad_norm": 0.33462053537368774, + "learning_rate": 0.00010936844408407016, + "loss": 0.2872, + "step": 35210 + }, + { + "epoch": 1.359898065562377, + "grad_norm": 0.9294387698173523, + "learning_rate": 0.00010934270306446839, + "loss": 0.2617, + "step": 35220 + }, + { + "epoch": 1.3602841808564037, + "grad_norm": 0.3169979453086853, + "learning_rate": 0.00010931696204486662, + "loss": 0.2942, + "step": 35230 + }, + { + "epoch": 1.3606702961504304, + "grad_norm": 2.1339616775512695, + "learning_rate": 0.00010929122102526482, + "loss": 0.4448, + "step": 35240 + }, + { + "epoch": 1.3610564114444572, + "grad_norm": 0.9430062770843506, + "learning_rate": 0.00010926548000566303, + "loss": 0.2051, + "step": 35250 + }, + { + "epoch": 1.3614425267384842, + "grad_norm": 3.1187360286712646, + "learning_rate": 0.00010923973898606123, + "loss": 0.2274, + "step": 35260 + }, + { + "epoch": 1.361828642032511, + "grad_norm": 1.4727579355239868, + "learning_rate": 0.00010921399796645946, + "loss": 0.3757, + "step": 35270 + }, + { + "epoch": 1.3622147573265377, + "grad_norm": 2.157560348510742, + "learning_rate": 0.00010918825694685768, + "loss": 0.3096, + "step": 35280 + }, + { + "epoch": 1.3626008726205645, + "grad_norm": 0.33457377552986145, + "learning_rate": 0.00010916251592725588, + "loss": 0.1489, + "step": 35290 + }, + { + "epoch": 1.3629869879145913, + "grad_norm": 0.9005904197692871, + "learning_rate": 0.00010913677490765411, + "loss": 0.1826, + "step": 35300 + }, + { + "epoch": 1.363373103208618, + "grad_norm": 2.1222829818725586, + "learning_rate": 0.00010911103388805231, + "loss": 0.1965, + "step": 35310 + }, + { + "epoch": 1.3637592185026448, + "grad_norm": 1.3881357908248901, + "learning_rate": 0.00010908529286845052, + "loss": 0.1791, + "step": 35320 + }, + { + "epoch": 1.3641453337966718, + "grad_norm": 1.7574503421783447, + "learning_rate": 0.00010905955184884872, + "loss": 0.3316, + "step": 35330 + }, + { + "epoch": 1.3645314490906983, + "grad_norm": 0.1967727392911911, + "learning_rate": 0.00010903381082924695, + "loss": 0.2331, + "step": 35340 + }, + { + "epoch": 1.3649175643847253, + "grad_norm": 0.8974360823631287, + "learning_rate": 0.00010900806980964518, + "loss": 0.2589, + "step": 35350 + }, + { + "epoch": 1.365303679678752, + "grad_norm": 2.0996744632720947, + "learning_rate": 0.00010898232879004338, + "loss": 0.3663, + "step": 35360 + }, + { + "epoch": 1.3656897949727789, + "grad_norm": 0.5678316354751587, + "learning_rate": 0.0001089565877704416, + "loss": 0.1729, + "step": 35370 + }, + { + "epoch": 1.3660759102668056, + "grad_norm": 2.3381874561309814, + "learning_rate": 0.0001089308467508398, + "loss": 0.1615, + "step": 35380 + }, + { + "epoch": 1.3664620255608324, + "grad_norm": 1.0276836156845093, + "learning_rate": 0.00010890510573123802, + "loss": 0.3359, + "step": 35390 + }, + { + "epoch": 1.3668481408548594, + "grad_norm": 2.4374940395355225, + "learning_rate": 0.00010887936471163622, + "loss": 0.2435, + "step": 35400 + }, + { + "epoch": 1.367234256148886, + "grad_norm": 0.45221665501594543, + "learning_rate": 0.00010885362369203444, + "loss": 0.2555, + "step": 35410 + }, + { + "epoch": 1.367620371442913, + "grad_norm": 2.608090400695801, + "learning_rate": 0.00010882788267243267, + "loss": 0.2465, + "step": 35420 + }, + { + "epoch": 1.3680064867369397, + "grad_norm": 1.4186642169952393, + "learning_rate": 0.00010880214165283087, + "loss": 0.1674, + "step": 35430 + }, + { + "epoch": 1.3683926020309665, + "grad_norm": 0.659479022026062, + "learning_rate": 0.0001087764006332291, + "loss": 0.2926, + "step": 35440 + }, + { + "epoch": 1.3687787173249932, + "grad_norm": 0.9219567179679871, + "learning_rate": 0.0001087506596136273, + "loss": 0.2001, + "step": 35450 + }, + { + "epoch": 1.36916483261902, + "grad_norm": 0.8070804476737976, + "learning_rate": 0.00010872491859402551, + "loss": 0.2178, + "step": 35460 + }, + { + "epoch": 1.3695509479130468, + "grad_norm": 2.9981069564819336, + "learning_rate": 0.00010869917757442374, + "loss": 0.3079, + "step": 35470 + }, + { + "epoch": 1.3699370632070735, + "grad_norm": 0.7891242504119873, + "learning_rate": 0.00010867343655482194, + "loss": 0.2765, + "step": 35480 + }, + { + "epoch": 1.3703231785011005, + "grad_norm": 1.448637843132019, + "learning_rate": 0.00010864769553522016, + "loss": 0.3521, + "step": 35490 + }, + { + "epoch": 1.3707092937951273, + "grad_norm": 0.07628043740987778, + "learning_rate": 0.00010862195451561836, + "loss": 0.2083, + "step": 35500 + }, + { + "epoch": 1.371095409089154, + "grad_norm": 0.7549735307693481, + "learning_rate": 0.00010859621349601659, + "loss": 0.2536, + "step": 35510 + }, + { + "epoch": 1.3714815243831808, + "grad_norm": 1.3548041582107544, + "learning_rate": 0.00010857047247641479, + "loss": 0.251, + "step": 35520 + }, + { + "epoch": 1.3718676396772076, + "grad_norm": 0.530010998249054, + "learning_rate": 0.000108544731456813, + "loss": 0.1917, + "step": 35530 + }, + { + "epoch": 1.3722537549712344, + "grad_norm": 0.4148992896080017, + "learning_rate": 0.00010851899043721123, + "loss": 0.335, + "step": 35540 + }, + { + "epoch": 1.3726398702652611, + "grad_norm": 1.5118776559829712, + "learning_rate": 0.00010849324941760943, + "loss": 0.2159, + "step": 35550 + }, + { + "epoch": 1.3730259855592881, + "grad_norm": 1.036889910697937, + "learning_rate": 0.00010846750839800766, + "loss": 0.2975, + "step": 35560 + }, + { + "epoch": 1.3734121008533147, + "grad_norm": 1.724263072013855, + "learning_rate": 0.00010844176737840586, + "loss": 0.1476, + "step": 35570 + }, + { + "epoch": 1.3737982161473417, + "grad_norm": 1.599007487297058, + "learning_rate": 0.00010841602635880408, + "loss": 0.2539, + "step": 35580 + }, + { + "epoch": 1.3741843314413684, + "grad_norm": 2.9119279384613037, + "learning_rate": 0.00010839028533920228, + "loss": 0.2688, + "step": 35590 + }, + { + "epoch": 1.3745704467353952, + "grad_norm": 1.8647874593734741, + "learning_rate": 0.00010836454431960051, + "loss": 0.4158, + "step": 35600 + }, + { + "epoch": 1.374956562029422, + "grad_norm": 3.925290822982788, + "learning_rate": 0.00010833880329999872, + "loss": 0.3333, + "step": 35610 + }, + { + "epoch": 1.3753426773234487, + "grad_norm": 0.7124634385108948, + "learning_rate": 0.00010831306228039692, + "loss": 0.1069, + "step": 35620 + }, + { + "epoch": 1.3757287926174757, + "grad_norm": 1.303579330444336, + "learning_rate": 0.00010828732126079515, + "loss": 0.2898, + "step": 35630 + }, + { + "epoch": 1.3761149079115023, + "grad_norm": 3.921804189682007, + "learning_rate": 0.00010826158024119335, + "loss": 0.4212, + "step": 35640 + }, + { + "epoch": 1.3765010232055293, + "grad_norm": 1.3194564580917358, + "learning_rate": 0.00010823583922159158, + "loss": 0.2771, + "step": 35650 + }, + { + "epoch": 1.376887138499556, + "grad_norm": 1.4237637519836426, + "learning_rate": 0.00010821009820198979, + "loss": 0.2463, + "step": 35660 + }, + { + "epoch": 1.3772732537935828, + "grad_norm": 1.8165888786315918, + "learning_rate": 0.000108184357182388, + "loss": 0.291, + "step": 35670 + }, + { + "epoch": 1.3776593690876096, + "grad_norm": 1.1056426763534546, + "learning_rate": 0.00010815861616278622, + "loss": 0.2525, + "step": 35680 + }, + { + "epoch": 1.3780454843816363, + "grad_norm": 1.483189582824707, + "learning_rate": 0.00010813287514318442, + "loss": 0.1569, + "step": 35690 + }, + { + "epoch": 1.378431599675663, + "grad_norm": 1.0666841268539429, + "learning_rate": 0.00010810713412358264, + "loss": 0.235, + "step": 35700 + }, + { + "epoch": 1.3788177149696899, + "grad_norm": 1.0299845933914185, + "learning_rate": 0.00010808139310398084, + "loss": 0.3892, + "step": 35710 + }, + { + "epoch": 1.3792038302637168, + "grad_norm": 2.3474409580230713, + "learning_rate": 0.00010805565208437907, + "loss": 0.3417, + "step": 35720 + }, + { + "epoch": 1.3795899455577436, + "grad_norm": 1.7456315755844116, + "learning_rate": 0.0001080299110647773, + "loss": 0.2538, + "step": 35730 + }, + { + "epoch": 1.3799760608517704, + "grad_norm": 2.866103410720825, + "learning_rate": 0.0001080041700451755, + "loss": 0.1619, + "step": 35740 + }, + { + "epoch": 1.3803621761457971, + "grad_norm": 0.29136407375335693, + "learning_rate": 0.00010797842902557371, + "loss": 0.2692, + "step": 35750 + }, + { + "epoch": 1.380748291439824, + "grad_norm": 0.8046161532402039, + "learning_rate": 0.00010795268800597191, + "loss": 0.1575, + "step": 35760 + }, + { + "epoch": 1.3811344067338507, + "grad_norm": 0.6451787352561951, + "learning_rate": 0.00010792694698637014, + "loss": 0.4914, + "step": 35770 + }, + { + "epoch": 1.3815205220278775, + "grad_norm": 0.7289161086082458, + "learning_rate": 0.00010790120596676836, + "loss": 0.1895, + "step": 35780 + }, + { + "epoch": 1.3819066373219044, + "grad_norm": 0.8300430178642273, + "learning_rate": 0.00010787546494716656, + "loss": 0.3663, + "step": 35790 + }, + { + "epoch": 1.382292752615931, + "grad_norm": 0.17713364958763123, + "learning_rate": 0.00010784972392756479, + "loss": 0.3189, + "step": 35800 + }, + { + "epoch": 1.382678867909958, + "grad_norm": 0.903222918510437, + "learning_rate": 0.00010782398290796299, + "loss": 0.1577, + "step": 35810 + }, + { + "epoch": 1.3830649832039847, + "grad_norm": 0.08617932349443436, + "learning_rate": 0.0001077982418883612, + "loss": 0.2872, + "step": 35820 + }, + { + "epoch": 1.3834510984980115, + "grad_norm": 1.9590895175933838, + "learning_rate": 0.0001077725008687594, + "loss": 0.2907, + "step": 35830 + }, + { + "epoch": 1.3838372137920383, + "grad_norm": 1.2515161037445068, + "learning_rate": 0.00010774675984915763, + "loss": 0.177, + "step": 35840 + }, + { + "epoch": 1.384223329086065, + "grad_norm": 1.6171292066574097, + "learning_rate": 0.00010772101882955586, + "loss": 0.2321, + "step": 35850 + }, + { + "epoch": 1.3846094443800918, + "grad_norm": 0.13681405782699585, + "learning_rate": 0.00010769527780995406, + "loss": 0.247, + "step": 35860 + }, + { + "epoch": 1.3849955596741186, + "grad_norm": 1.1949968338012695, + "learning_rate": 0.00010766953679035228, + "loss": 0.2978, + "step": 35870 + }, + { + "epoch": 1.3853816749681456, + "grad_norm": 0.17001692950725555, + "learning_rate": 0.00010764379577075048, + "loss": 0.3241, + "step": 35880 + }, + { + "epoch": 1.3857677902621723, + "grad_norm": 0.8227952122688293, + "learning_rate": 0.0001076180547511487, + "loss": 0.3499, + "step": 35890 + }, + { + "epoch": 1.386153905556199, + "grad_norm": 1.4185482263565063, + "learning_rate": 0.0001075923137315469, + "loss": 0.3109, + "step": 35900 + }, + { + "epoch": 1.3865400208502259, + "grad_norm": 0.9533351063728333, + "learning_rate": 0.00010756657271194512, + "loss": 0.24, + "step": 35910 + }, + { + "epoch": 1.3869261361442526, + "grad_norm": 1.077789306640625, + "learning_rate": 0.00010754083169234335, + "loss": 0.2662, + "step": 35920 + }, + { + "epoch": 1.3873122514382794, + "grad_norm": 1.3528363704681396, + "learning_rate": 0.00010751509067274155, + "loss": 0.1623, + "step": 35930 + }, + { + "epoch": 1.3876983667323062, + "grad_norm": 0.25122806429862976, + "learning_rate": 0.00010748934965313978, + "loss": 0.1639, + "step": 35940 + }, + { + "epoch": 1.3880844820263332, + "grad_norm": 0.9446159601211548, + "learning_rate": 0.00010746360863353798, + "loss": 0.2035, + "step": 35950 + }, + { + "epoch": 1.38847059732036, + "grad_norm": 1.2258719205856323, + "learning_rate": 0.00010743786761393619, + "loss": 0.4247, + "step": 35960 + }, + { + "epoch": 1.3888567126143867, + "grad_norm": 3.117729663848877, + "learning_rate": 0.00010741212659433442, + "loss": 0.2569, + "step": 35970 + }, + { + "epoch": 1.3892428279084135, + "grad_norm": 0.6333123445510864, + "learning_rate": 0.00010738638557473262, + "loss": 0.1843, + "step": 35980 + }, + { + "epoch": 1.3896289432024402, + "grad_norm": 1.49360191822052, + "learning_rate": 0.00010736064455513084, + "loss": 0.2318, + "step": 35990 + }, + { + "epoch": 1.390015058496467, + "grad_norm": 3.9082753658294678, + "learning_rate": 0.00010733490353552904, + "loss": 0.2037, + "step": 36000 + }, + { + "epoch": 1.3904011737904938, + "grad_norm": 0.5687323808670044, + "learning_rate": 0.00010730916251592727, + "loss": 0.0871, + "step": 36010 + }, + { + "epoch": 1.3907872890845208, + "grad_norm": 0.3480868339538574, + "learning_rate": 0.00010728342149632547, + "loss": 0.4009, + "step": 36020 + }, + { + "epoch": 1.3911734043785473, + "grad_norm": 1.325042486190796, + "learning_rate": 0.00010725768047672368, + "loss": 0.5143, + "step": 36030 + }, + { + "epoch": 1.3915595196725743, + "grad_norm": 2.114786386489868, + "learning_rate": 0.00010723193945712191, + "loss": 0.2944, + "step": 36040 + }, + { + "epoch": 1.391945634966601, + "grad_norm": 1.716272234916687, + "learning_rate": 0.00010720619843752011, + "loss": 0.3793, + "step": 36050 + }, + { + "epoch": 1.3923317502606278, + "grad_norm": 2.057535171508789, + "learning_rate": 0.00010718045741791834, + "loss": 0.199, + "step": 36060 + }, + { + "epoch": 1.3927178655546546, + "grad_norm": 2.371248483657837, + "learning_rate": 0.00010715471639831653, + "loss": 0.3276, + "step": 36070 + }, + { + "epoch": 1.3931039808486814, + "grad_norm": 0.1784186065196991, + "learning_rate": 0.00010712897537871476, + "loss": 0.0922, + "step": 36080 + }, + { + "epoch": 1.3934900961427081, + "grad_norm": 0.7923040390014648, + "learning_rate": 0.00010710323435911298, + "loss": 0.1599, + "step": 36090 + }, + { + "epoch": 1.393876211436735, + "grad_norm": 1.9013831615447998, + "learning_rate": 0.00010707749333951117, + "loss": 0.4997, + "step": 36100 + }, + { + "epoch": 1.394262326730762, + "grad_norm": 1.8659415245056152, + "learning_rate": 0.0001070517523199094, + "loss": 0.1851, + "step": 36110 + }, + { + "epoch": 1.3946484420247887, + "grad_norm": 2.5775375366210938, + "learning_rate": 0.0001070260113003076, + "loss": 0.2384, + "step": 36120 + }, + { + "epoch": 1.3950345573188154, + "grad_norm": 0.21943879127502441, + "learning_rate": 0.00010700027028070583, + "loss": 0.3996, + "step": 36130 + }, + { + "epoch": 1.3954206726128422, + "grad_norm": 1.1734743118286133, + "learning_rate": 0.00010697452926110403, + "loss": 0.1737, + "step": 36140 + }, + { + "epoch": 1.395806787906869, + "grad_norm": 0.697695791721344, + "learning_rate": 0.00010694878824150225, + "loss": 0.2778, + "step": 36150 + }, + { + "epoch": 1.3961929032008957, + "grad_norm": 3.2881579399108887, + "learning_rate": 0.00010692304722190047, + "loss": 0.3198, + "step": 36160 + }, + { + "epoch": 1.3965790184949225, + "grad_norm": 0.1592467725276947, + "learning_rate": 0.00010689730620229868, + "loss": 0.2591, + "step": 36170 + }, + { + "epoch": 1.3969651337889495, + "grad_norm": 0.4579029083251953, + "learning_rate": 0.0001068715651826969, + "loss": 0.2984, + "step": 36180 + }, + { + "epoch": 1.3973512490829763, + "grad_norm": 0.4286015033721924, + "learning_rate": 0.0001068458241630951, + "loss": 0.4098, + "step": 36190 + }, + { + "epoch": 1.397737364377003, + "grad_norm": 1.7824127674102783, + "learning_rate": 0.00010682008314349332, + "loss": 0.2446, + "step": 36200 + }, + { + "epoch": 1.3981234796710298, + "grad_norm": 0.8584449887275696, + "learning_rate": 0.00010679434212389152, + "loss": 0.1415, + "step": 36210 + }, + { + "epoch": 1.3985095949650566, + "grad_norm": 1.1699339151382446, + "learning_rate": 0.00010676860110428975, + "loss": 0.1706, + "step": 36220 + }, + { + "epoch": 1.3988957102590833, + "grad_norm": 2.615877389907837, + "learning_rate": 0.00010674286008468796, + "loss": 0.2605, + "step": 36230 + }, + { + "epoch": 1.39928182555311, + "grad_norm": 2.182037591934204, + "learning_rate": 0.00010671711906508617, + "loss": 0.2067, + "step": 36240 + }, + { + "epoch": 1.399667940847137, + "grad_norm": 2.183263063430786, + "learning_rate": 0.00010669137804548439, + "loss": 0.4704, + "step": 36250 + }, + { + "epoch": 1.4000540561411636, + "grad_norm": 3.505791187286377, + "learning_rate": 0.00010666563702588259, + "loss": 0.3385, + "step": 36260 + }, + { + "epoch": 1.4004401714351906, + "grad_norm": 1.2262030839920044, + "learning_rate": 0.00010663989600628081, + "loss": 0.3999, + "step": 36270 + }, + { + "epoch": 1.4008262867292174, + "grad_norm": 2.4024577140808105, + "learning_rate": 0.00010661415498667904, + "loss": 0.1984, + "step": 36280 + }, + { + "epoch": 1.4012124020232442, + "grad_norm": 0.4166090786457062, + "learning_rate": 0.00010658841396707724, + "loss": 0.2448, + "step": 36290 + }, + { + "epoch": 1.401598517317271, + "grad_norm": 0.422590047121048, + "learning_rate": 0.00010656267294747547, + "loss": 0.2464, + "step": 36300 + }, + { + "epoch": 1.4019846326112977, + "grad_norm": 2.287503480911255, + "learning_rate": 0.00010653693192787367, + "loss": 0.1621, + "step": 36310 + }, + { + "epoch": 1.4023707479053245, + "grad_norm": 0.8126110434532166, + "learning_rate": 0.00010651119090827188, + "loss": 0.2097, + "step": 36320 + }, + { + "epoch": 1.4027568631993512, + "grad_norm": 0.683016836643219, + "learning_rate": 0.00010648544988867008, + "loss": 0.1512, + "step": 36330 + }, + { + "epoch": 1.4031429784933782, + "grad_norm": 1.5477893352508545, + "learning_rate": 0.00010645970886906831, + "loss": 0.2169, + "step": 36340 + }, + { + "epoch": 1.403529093787405, + "grad_norm": 2.183166265487671, + "learning_rate": 0.00010643396784946653, + "loss": 0.4307, + "step": 36350 + }, + { + "epoch": 1.4039152090814317, + "grad_norm": 1.5782747268676758, + "learning_rate": 0.00010640822682986473, + "loss": 0.291, + "step": 36360 + }, + { + "epoch": 1.4043013243754585, + "grad_norm": 1.2823392152786255, + "learning_rate": 0.00010638248581026296, + "loss": 0.315, + "step": 36370 + }, + { + "epoch": 1.4046874396694853, + "grad_norm": 2.0126500129699707, + "learning_rate": 0.00010635674479066116, + "loss": 0.2565, + "step": 36380 + }, + { + "epoch": 1.405073554963512, + "grad_norm": 2.490217447280884, + "learning_rate": 0.00010633100377105937, + "loss": 0.2665, + "step": 36390 + }, + { + "epoch": 1.4054596702575388, + "grad_norm": 1.2054855823516846, + "learning_rate": 0.00010630526275145757, + "loss": 0.5455, + "step": 36400 + }, + { + "epoch": 1.4058457855515658, + "grad_norm": 0.7968757748603821, + "learning_rate": 0.0001062795217318558, + "loss": 0.333, + "step": 36410 + }, + { + "epoch": 1.4062319008455926, + "grad_norm": 1.1027718782424927, + "learning_rate": 0.00010625378071225403, + "loss": 0.2019, + "step": 36420 + }, + { + "epoch": 1.4066180161396193, + "grad_norm": 1.987302541732788, + "learning_rate": 0.00010622803969265223, + "loss": 0.3159, + "step": 36430 + }, + { + "epoch": 1.407004131433646, + "grad_norm": 1.5426512956619263, + "learning_rate": 0.00010620229867305045, + "loss": 0.3759, + "step": 36440 + }, + { + "epoch": 1.4073902467276729, + "grad_norm": 0.5228156447410583, + "learning_rate": 0.00010617655765344865, + "loss": 0.2465, + "step": 36450 + }, + { + "epoch": 1.4077763620216996, + "grad_norm": 0.46890121698379517, + "learning_rate": 0.00010615081663384687, + "loss": 0.302, + "step": 36460 + }, + { + "epoch": 1.4081624773157264, + "grad_norm": 2.1506495475769043, + "learning_rate": 0.0001061250756142451, + "loss": 0.2569, + "step": 36470 + }, + { + "epoch": 1.4085485926097534, + "grad_norm": 2.307468891143799, + "learning_rate": 0.0001060993345946433, + "loss": 0.2009, + "step": 36480 + }, + { + "epoch": 1.40893470790378, + "grad_norm": 0.07033026963472366, + "learning_rate": 0.00010607359357504152, + "loss": 0.1728, + "step": 36490 + }, + { + "epoch": 1.409320823197807, + "grad_norm": 0.3262972831726074, + "learning_rate": 0.00010604785255543972, + "loss": 0.2905, + "step": 36500 + }, + { + "epoch": 1.4097069384918337, + "grad_norm": 0.755646824836731, + "learning_rate": 0.00010602211153583795, + "loss": 0.3287, + "step": 36510 + }, + { + "epoch": 1.4100930537858605, + "grad_norm": 1.1180161237716675, + "learning_rate": 0.00010599637051623615, + "loss": 0.2743, + "step": 36520 + }, + { + "epoch": 1.4104791690798872, + "grad_norm": 1.2358392477035522, + "learning_rate": 0.00010597062949663436, + "loss": 0.2672, + "step": 36530 + }, + { + "epoch": 1.410865284373914, + "grad_norm": 1.4188988208770752, + "learning_rate": 0.00010594488847703259, + "loss": 0.2552, + "step": 36540 + }, + { + "epoch": 1.4112513996679408, + "grad_norm": 1.4727978706359863, + "learning_rate": 0.00010591914745743079, + "loss": 0.2256, + "step": 36550 + }, + { + "epoch": 1.4116375149619675, + "grad_norm": 0.08973213285207748, + "learning_rate": 0.00010589340643782901, + "loss": 0.224, + "step": 36560 + }, + { + "epoch": 1.4120236302559945, + "grad_norm": 0.9915102124214172, + "learning_rate": 0.00010586766541822721, + "loss": 0.195, + "step": 36570 + }, + { + "epoch": 1.4124097455500213, + "grad_norm": 0.8524800539016724, + "learning_rate": 0.00010584192439862544, + "loss": 0.1492, + "step": 36580 + }, + { + "epoch": 1.412795860844048, + "grad_norm": 1.8414466381072998, + "learning_rate": 0.00010581618337902365, + "loss": 0.3412, + "step": 36590 + }, + { + "epoch": 1.4131819761380748, + "grad_norm": 2.596547842025757, + "learning_rate": 0.00010579044235942185, + "loss": 0.3469, + "step": 36600 + }, + { + "epoch": 1.4135680914321016, + "grad_norm": 0.5367813110351562, + "learning_rate": 0.00010576470133982008, + "loss": 0.2274, + "step": 36610 + }, + { + "epoch": 1.4139542067261284, + "grad_norm": 2.70858097076416, + "learning_rate": 0.00010573896032021828, + "loss": 0.171, + "step": 36620 + }, + { + "epoch": 1.4143403220201551, + "grad_norm": 1.7077667713165283, + "learning_rate": 0.00010571321930061651, + "loss": 0.2638, + "step": 36630 + }, + { + "epoch": 1.4147264373141821, + "grad_norm": 0.9189953804016113, + "learning_rate": 0.00010568747828101471, + "loss": 0.2283, + "step": 36640 + }, + { + "epoch": 1.4151125526082087, + "grad_norm": 2.0240087509155273, + "learning_rate": 0.00010566173726141293, + "loss": 0.3166, + "step": 36650 + }, + { + "epoch": 1.4154986679022357, + "grad_norm": 3.3304507732391357, + "learning_rate": 0.00010563599624181115, + "loss": 0.3046, + "step": 36660 + }, + { + "epoch": 1.4158847831962624, + "grad_norm": 1.955929160118103, + "learning_rate": 0.00010561025522220935, + "loss": 0.2653, + "step": 36670 + }, + { + "epoch": 1.4162708984902892, + "grad_norm": 2.2465381622314453, + "learning_rate": 0.00010558451420260757, + "loss": 0.3119, + "step": 36680 + }, + { + "epoch": 1.416657013784316, + "grad_norm": 2.108614921569824, + "learning_rate": 0.00010555877318300577, + "loss": 0.2872, + "step": 36690 + }, + { + "epoch": 1.4170431290783427, + "grad_norm": 1.35493004322052, + "learning_rate": 0.000105533032163404, + "loss": 0.1491, + "step": 36700 + }, + { + "epoch": 1.4174292443723697, + "grad_norm": 0.9102393984794617, + "learning_rate": 0.0001055072911438022, + "loss": 0.1565, + "step": 36710 + }, + { + "epoch": 1.4178153596663963, + "grad_norm": 1.7878345251083374, + "learning_rate": 0.00010548155012420043, + "loss": 0.4864, + "step": 36720 + }, + { + "epoch": 1.4182014749604233, + "grad_norm": 1.4333637952804565, + "learning_rate": 0.00010545580910459864, + "loss": 0.3246, + "step": 36730 + }, + { + "epoch": 1.41858759025445, + "grad_norm": 0.0966360941529274, + "learning_rate": 0.00010543006808499684, + "loss": 0.1758, + "step": 36740 + }, + { + "epoch": 1.4189737055484768, + "grad_norm": 0.30043545365333557, + "learning_rate": 0.00010540432706539507, + "loss": 0.1811, + "step": 36750 + }, + { + "epoch": 1.4193598208425036, + "grad_norm": 1.7705951929092407, + "learning_rate": 0.00010537858604579327, + "loss": 0.1824, + "step": 36760 + }, + { + "epoch": 1.4197459361365303, + "grad_norm": 1.2025195360183716, + "learning_rate": 0.0001053528450261915, + "loss": 0.1463, + "step": 36770 + }, + { + "epoch": 1.420132051430557, + "grad_norm": 0.3154304325580597, + "learning_rate": 0.00010532710400658972, + "loss": 0.1817, + "step": 36780 + }, + { + "epoch": 1.4205181667245839, + "grad_norm": 3.392331838607788, + "learning_rate": 0.00010530136298698792, + "loss": 0.2938, + "step": 36790 + }, + { + "epoch": 1.4209042820186109, + "grad_norm": 0.7256132364273071, + "learning_rate": 0.00010527562196738613, + "loss": 0.28, + "step": 36800 + }, + { + "epoch": 1.4212903973126376, + "grad_norm": 1.9007991552352905, + "learning_rate": 0.00010524988094778435, + "loss": 0.2306, + "step": 36810 + }, + { + "epoch": 1.4216765126066644, + "grad_norm": 3.21189546585083, + "learning_rate": 0.00010522413992818256, + "loss": 0.2945, + "step": 36820 + }, + { + "epoch": 1.4220626279006912, + "grad_norm": 0.20476600527763367, + "learning_rate": 0.00010519839890858076, + "loss": 0.3129, + "step": 36830 + }, + { + "epoch": 1.422448743194718, + "grad_norm": 1.4225107431411743, + "learning_rate": 0.00010517265788897899, + "loss": 0.2262, + "step": 36840 + }, + { + "epoch": 1.4228348584887447, + "grad_norm": 1.203728437423706, + "learning_rate": 0.00010514691686937721, + "loss": 0.176, + "step": 36850 + }, + { + "epoch": 1.4232209737827715, + "grad_norm": 0.3001759648323059, + "learning_rate": 0.00010512117584977541, + "loss": 0.218, + "step": 36860 + }, + { + "epoch": 1.4236070890767984, + "grad_norm": 2.3143389225006104, + "learning_rate": 0.00010509543483017363, + "loss": 0.1562, + "step": 36870 + }, + { + "epoch": 1.423993204370825, + "grad_norm": 0.5087364912033081, + "learning_rate": 0.00010506969381057184, + "loss": 0.1207, + "step": 36880 + }, + { + "epoch": 1.424379319664852, + "grad_norm": 1.6521960496902466, + "learning_rate": 0.00010504395279097005, + "loss": 0.4756, + "step": 36890 + }, + { + "epoch": 1.4247654349588788, + "grad_norm": 0.09236706793308258, + "learning_rate": 0.00010501821177136825, + "loss": 0.1755, + "step": 36900 + }, + { + "epoch": 1.4251515502529055, + "grad_norm": 0.2143094390630722, + "learning_rate": 0.00010499247075176648, + "loss": 0.2126, + "step": 36910 + }, + { + "epoch": 1.4255376655469323, + "grad_norm": 1.210170865058899, + "learning_rate": 0.00010496672973216471, + "loss": 0.1391, + "step": 36920 + }, + { + "epoch": 1.425923780840959, + "grad_norm": 0.6248244047164917, + "learning_rate": 0.00010494098871256291, + "loss": 0.2898, + "step": 36930 + }, + { + "epoch": 1.426309896134986, + "grad_norm": 2.1100337505340576, + "learning_rate": 0.00010491524769296113, + "loss": 0.1198, + "step": 36940 + }, + { + "epoch": 1.4266960114290126, + "grad_norm": 2.5673348903656006, + "learning_rate": 0.00010488950667335933, + "loss": 0.2284, + "step": 36950 + }, + { + "epoch": 1.4270821267230396, + "grad_norm": 1.868195652961731, + "learning_rate": 0.00010486376565375755, + "loss": 0.3738, + "step": 36960 + }, + { + "epoch": 1.4274682420170663, + "grad_norm": 1.0951671600341797, + "learning_rate": 0.00010483802463415577, + "loss": 0.1424, + "step": 36970 + }, + { + "epoch": 1.4278543573110931, + "grad_norm": 0.4791143536567688, + "learning_rate": 0.00010481228361455397, + "loss": 0.1445, + "step": 36980 + }, + { + "epoch": 1.4282404726051199, + "grad_norm": 3.2477540969848633, + "learning_rate": 0.0001047865425949522, + "loss": 0.2676, + "step": 36990 + }, + { + "epoch": 1.4286265878991466, + "grad_norm": 0.8082342147827148, + "learning_rate": 0.0001047608015753504, + "loss": 0.2295, + "step": 37000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.253144195072e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..548ac030e229bf4cfddca9f329208c397f599148 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b6e4397795b6a05cb45aa07f091937fd34816ba953a828f7c91067d62067878 +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..85c120c8e17efbcffde2b7384c8342f1beb9c82f --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d11ff63a28334a70223574ce52a378fc24a23a5fb1dbd904b02c3834a27a25b2 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..7b9491294b29b55f6494707be11f1d0096023419 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..5bf4071ba148fff946a426d47d0e971e96da793a Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..892261bb6292f447b0120f028ba573e225bfde60 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/trainer_state.json @@ -0,0 +1,26284 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.4479323526004866, + "eval_steps": 500, + "global_step": 37500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + }, + { + "epoch": 1.274566585582455, + "grad_norm": 0.5019121766090393, + "learning_rate": 0.0001150314683964632, + "loss": 0.1857, + "step": 33010 + }, + { + "epoch": 1.2749527008764816, + "grad_norm": 0.258952260017395, + "learning_rate": 0.0001150057273768614, + "loss": 0.11, + "step": 33020 + }, + { + "epoch": 1.2753388161705086, + "grad_norm": 0.8540348410606384, + "learning_rate": 0.00011497998635725962, + "loss": 0.1852, + "step": 33030 + }, + { + "epoch": 1.2757249314645354, + "grad_norm": 0.08689398318529129, + "learning_rate": 0.00011495424533765783, + "loss": 0.2225, + "step": 33040 + }, + { + "epoch": 1.2761110467585621, + "grad_norm": 0.42253822088241577, + "learning_rate": 0.00011492850431805605, + "loss": 0.0751, + "step": 33050 + }, + { + "epoch": 1.276497162052589, + "grad_norm": 1.2964017391204834, + "learning_rate": 0.00011490276329845426, + "loss": 0.2384, + "step": 33060 + }, + { + "epoch": 1.2768832773466157, + "grad_norm": 0.5337836146354675, + "learning_rate": 0.00011487702227885246, + "loss": 0.1415, + "step": 33070 + }, + { + "epoch": 1.2772693926406424, + "grad_norm": 2.7771682739257812, + "learning_rate": 0.00011485128125925069, + "loss": 0.378, + "step": 33080 + }, + { + "epoch": 1.2776555079346692, + "grad_norm": 1.5107232332229614, + "learning_rate": 0.00011482554023964889, + "loss": 0.2482, + "step": 33090 + }, + { + "epoch": 1.2780416232286962, + "grad_norm": 0.6885499358177185, + "learning_rate": 0.00011479979922004711, + "loss": 0.2122, + "step": 33100 + }, + { + "epoch": 1.278427738522723, + "grad_norm": 0.9016557335853577, + "learning_rate": 0.00011477405820044533, + "loss": 0.2841, + "step": 33110 + }, + { + "epoch": 1.2788138538167497, + "grad_norm": 1.9532525539398193, + "learning_rate": 0.00011474831718084354, + "loss": 0.2281, + "step": 33120 + }, + { + "epoch": 1.2791999691107765, + "grad_norm": 2.1078782081604004, + "learning_rate": 0.00011472257616124175, + "loss": 0.2836, + "step": 33130 + }, + { + "epoch": 1.2795860844048033, + "grad_norm": 0.19830390810966492, + "learning_rate": 0.00011469683514163995, + "loss": 0.388, + "step": 33140 + }, + { + "epoch": 1.27997219969883, + "grad_norm": 0.17538850009441376, + "learning_rate": 0.00011467109412203818, + "loss": 0.3274, + "step": 33150 + }, + { + "epoch": 1.2803583149928568, + "grad_norm": 0.7402139902114868, + "learning_rate": 0.00011464535310243641, + "loss": 0.1979, + "step": 33160 + }, + { + "epoch": 1.2807444302868838, + "grad_norm": 0.2097146362066269, + "learning_rate": 0.00011461961208283461, + "loss": 0.2464, + "step": 33170 + }, + { + "epoch": 1.2811305455809103, + "grad_norm": 1.2441083192825317, + "learning_rate": 0.00011459387106323283, + "loss": 0.266, + "step": 33180 + }, + { + "epoch": 1.2815166608749373, + "grad_norm": 2.518852710723877, + "learning_rate": 0.00011456813004363103, + "loss": 0.253, + "step": 33190 + }, + { + "epoch": 1.281902776168964, + "grad_norm": 0.8078998327255249, + "learning_rate": 0.00011454238902402925, + "loss": 0.2361, + "step": 33200 + }, + { + "epoch": 1.2822888914629909, + "grad_norm": 1.2297371625900269, + "learning_rate": 0.00011451664800442745, + "loss": 0.1974, + "step": 33210 + }, + { + "epoch": 1.2826750067570176, + "grad_norm": 0.4303855895996094, + "learning_rate": 0.00011449090698482567, + "loss": 0.3563, + "step": 33220 + }, + { + "epoch": 1.2830611220510444, + "grad_norm": 1.3215210437774658, + "learning_rate": 0.0001144651659652239, + "loss": 0.2818, + "step": 33230 + }, + { + "epoch": 1.2834472373450712, + "grad_norm": 1.546265959739685, + "learning_rate": 0.0001144394249456221, + "loss": 0.5778, + "step": 33240 + }, + { + "epoch": 1.283833352639098, + "grad_norm": 0.8895953297615051, + "learning_rate": 0.00011441368392602033, + "loss": 0.2485, + "step": 33250 + }, + { + "epoch": 1.284219467933125, + "grad_norm": 0.7534870505332947, + "learning_rate": 0.00011438794290641853, + "loss": 0.2825, + "step": 33260 + }, + { + "epoch": 1.2846055832271517, + "grad_norm": 0.052820973098278046, + "learning_rate": 0.00011436220188681674, + "loss": 0.2191, + "step": 33270 + }, + { + "epoch": 1.2849916985211784, + "grad_norm": 0.9264475107192993, + "learning_rate": 0.00011433646086721494, + "loss": 0.181, + "step": 33280 + }, + { + "epoch": 1.2853778138152052, + "grad_norm": 0.2128441333770752, + "learning_rate": 0.00011431071984761317, + "loss": 0.1819, + "step": 33290 + }, + { + "epoch": 1.285763929109232, + "grad_norm": 0.5400950312614441, + "learning_rate": 0.0001142849788280114, + "loss": 0.4316, + "step": 33300 + }, + { + "epoch": 1.2861500444032588, + "grad_norm": 0.8033271431922913, + "learning_rate": 0.00011425923780840959, + "loss": 0.2146, + "step": 33310 + }, + { + "epoch": 1.2865361596972855, + "grad_norm": 2.012575149536133, + "learning_rate": 0.00011423349678880782, + "loss": 0.4335, + "step": 33320 + }, + { + "epoch": 1.2869222749913125, + "grad_norm": 0.7352376580238342, + "learning_rate": 0.00011420775576920602, + "loss": 0.2124, + "step": 33330 + }, + { + "epoch": 1.2873083902853393, + "grad_norm": 0.769036591053009, + "learning_rate": 0.00011418201474960423, + "loss": 0.3602, + "step": 33340 + }, + { + "epoch": 1.287694505579366, + "grad_norm": 0.250592976808548, + "learning_rate": 0.00011415627373000246, + "loss": 0.1692, + "step": 33350 + }, + { + "epoch": 1.2880806208733928, + "grad_norm": 2.43820858001709, + "learning_rate": 0.00011413053271040066, + "loss": 0.2777, + "step": 33360 + }, + { + "epoch": 1.2884667361674196, + "grad_norm": 1.3179954290390015, + "learning_rate": 0.00011410479169079889, + "loss": 0.1794, + "step": 33370 + }, + { + "epoch": 1.2888528514614463, + "grad_norm": 1.0040466785430908, + "learning_rate": 0.00011407905067119709, + "loss": 0.3037, + "step": 33380 + }, + { + "epoch": 1.2892389667554731, + "grad_norm": 5.296288013458252, + "learning_rate": 0.00011405330965159531, + "loss": 0.2904, + "step": 33390 + }, + { + "epoch": 1.2896250820495, + "grad_norm": 0.4267273247241974, + "learning_rate": 0.00011402756863199351, + "loss": 0.2263, + "step": 33400 + }, + { + "epoch": 1.2900111973435266, + "grad_norm": 0.8817713260650635, + "learning_rate": 0.00011400182761239173, + "loss": 0.2715, + "step": 33410 + }, + { + "epoch": 1.2903973126375536, + "grad_norm": 2.7891275882720947, + "learning_rate": 0.00011397608659278995, + "loss": 0.2781, + "step": 33420 + }, + { + "epoch": 1.2907834279315804, + "grad_norm": 0.3013952374458313, + "learning_rate": 0.00011395034557318815, + "loss": 0.2563, + "step": 33430 + }, + { + "epoch": 1.2911695432256072, + "grad_norm": 1.766413927078247, + "learning_rate": 0.00011392460455358638, + "loss": 0.1854, + "step": 33440 + }, + { + "epoch": 1.291555658519634, + "grad_norm": 0.25331103801727295, + "learning_rate": 0.00011389886353398458, + "loss": 0.1206, + "step": 33450 + }, + { + "epoch": 1.2919417738136607, + "grad_norm": 0.036400288343429565, + "learning_rate": 0.0001138731225143828, + "loss": 0.4707, + "step": 33460 + }, + { + "epoch": 1.2923278891076875, + "grad_norm": 1.5620888471603394, + "learning_rate": 0.00011384738149478102, + "loss": 0.3312, + "step": 33470 + }, + { + "epoch": 1.2927140044017142, + "grad_norm": 0.6670392155647278, + "learning_rate": 0.00011382164047517922, + "loss": 0.2341, + "step": 33480 + }, + { + "epoch": 1.2931001196957412, + "grad_norm": 2.3108737468719482, + "learning_rate": 0.00011379589945557745, + "loss": 0.3843, + "step": 33490 + }, + { + "epoch": 1.293486234989768, + "grad_norm": 0.8025147318840027, + "learning_rate": 0.00011377015843597565, + "loss": 0.1982, + "step": 33500 + }, + { + "epoch": 1.2938723502837948, + "grad_norm": 1.7835719585418701, + "learning_rate": 0.00011374441741637387, + "loss": 0.3285, + "step": 33510 + }, + { + "epoch": 1.2942584655778215, + "grad_norm": 2.041508913040161, + "learning_rate": 0.00011371867639677207, + "loss": 0.2044, + "step": 33520 + }, + { + "epoch": 1.2946445808718483, + "grad_norm": 1.103378415107727, + "learning_rate": 0.0001136929353771703, + "loss": 0.1682, + "step": 33530 + }, + { + "epoch": 1.295030696165875, + "grad_norm": 0.057376351207494736, + "learning_rate": 0.00011366719435756851, + "loss": 0.1642, + "step": 33540 + }, + { + "epoch": 1.2954168114599018, + "grad_norm": 0.6539410948753357, + "learning_rate": 0.00011364145333796671, + "loss": 0.1549, + "step": 33550 + }, + { + "epoch": 1.2958029267539288, + "grad_norm": 1.250543236732483, + "learning_rate": 0.00011361571231836494, + "loss": 0.3764, + "step": 33560 + }, + { + "epoch": 1.2961890420479556, + "grad_norm": 0.23697887361049652, + "learning_rate": 0.00011358997129876314, + "loss": 0.3999, + "step": 33570 + }, + { + "epoch": 1.2965751573419824, + "grad_norm": 0.9318505525588989, + "learning_rate": 0.00011356423027916137, + "loss": 0.4156, + "step": 33580 + }, + { + "epoch": 1.2969612726360091, + "grad_norm": 1.3910777568817139, + "learning_rate": 0.00011353848925955957, + "loss": 0.3455, + "step": 33590 + }, + { + "epoch": 1.297347387930036, + "grad_norm": 1.6764451265335083, + "learning_rate": 0.00011351274823995779, + "loss": 0.1884, + "step": 33600 + }, + { + "epoch": 1.2977335032240627, + "grad_norm": 0.9300051927566528, + "learning_rate": 0.000113487007220356, + "loss": 0.119, + "step": 33610 + }, + { + "epoch": 1.2981196185180894, + "grad_norm": 2.447462558746338, + "learning_rate": 0.00011346126620075422, + "loss": 0.4403, + "step": 33620 + }, + { + "epoch": 1.2985057338121164, + "grad_norm": 1.216407060623169, + "learning_rate": 0.00011343552518115243, + "loss": 0.2415, + "step": 33630 + }, + { + "epoch": 1.298891849106143, + "grad_norm": 2.968648910522461, + "learning_rate": 0.00011340978416155063, + "loss": 0.2899, + "step": 33640 + }, + { + "epoch": 1.29927796440017, + "grad_norm": 0.6649970412254333, + "learning_rate": 0.00011338404314194886, + "loss": 0.3809, + "step": 33650 + }, + { + "epoch": 1.2996640796941967, + "grad_norm": 1.7277917861938477, + "learning_rate": 0.00011335830212234709, + "loss": 0.3308, + "step": 33660 + }, + { + "epoch": 1.3000501949882235, + "grad_norm": 1.3269709348678589, + "learning_rate": 0.00011333256110274529, + "loss": 0.3682, + "step": 33670 + }, + { + "epoch": 1.3004363102822503, + "grad_norm": 0.20609407126903534, + "learning_rate": 0.0001133068200831435, + "loss": 0.1379, + "step": 33680 + }, + { + "epoch": 1.300822425576277, + "grad_norm": 0.6592215299606323, + "learning_rate": 0.00011328107906354171, + "loss": 0.2746, + "step": 33690 + }, + { + "epoch": 1.3012085408703038, + "grad_norm": 1.903635859489441, + "learning_rate": 0.00011325533804393993, + "loss": 0.4729, + "step": 33700 + }, + { + "epoch": 1.3015946561643306, + "grad_norm": 0.8432504534721375, + "learning_rate": 0.00011322959702433813, + "loss": 0.2835, + "step": 33710 + }, + { + "epoch": 1.3019807714583576, + "grad_norm": 0.9862542152404785, + "learning_rate": 0.00011320385600473635, + "loss": 0.1823, + "step": 33720 + }, + { + "epoch": 1.3023668867523843, + "grad_norm": 3.845738649368286, + "learning_rate": 0.00011317811498513458, + "loss": 0.2542, + "step": 33730 + }, + { + "epoch": 1.302753002046411, + "grad_norm": 0.6317747235298157, + "learning_rate": 0.00011315237396553278, + "loss": 0.22, + "step": 33740 + }, + { + "epoch": 1.3031391173404379, + "grad_norm": 2.5221354961395264, + "learning_rate": 0.000113126632945931, + "loss": 0.2253, + "step": 33750 + }, + { + "epoch": 1.3035252326344646, + "grad_norm": 1.3326247930526733, + "learning_rate": 0.0001131008919263292, + "loss": 0.2021, + "step": 33760 + }, + { + "epoch": 1.3039113479284914, + "grad_norm": 1.148047685623169, + "learning_rate": 0.00011307515090672742, + "loss": 0.3987, + "step": 33770 + }, + { + "epoch": 1.3042974632225182, + "grad_norm": 0.19721268117427826, + "learning_rate": 0.00011304940988712562, + "loss": 0.2642, + "step": 33780 + }, + { + "epoch": 1.3046835785165452, + "grad_norm": 1.4060617685317993, + "learning_rate": 0.00011302366886752385, + "loss": 0.2736, + "step": 33790 + }, + { + "epoch": 1.305069693810572, + "grad_norm": 1.0736548900604248, + "learning_rate": 0.00011299792784792207, + "loss": 0.2448, + "step": 33800 + }, + { + "epoch": 1.3054558091045987, + "grad_norm": 4.352476119995117, + "learning_rate": 0.00011297218682832027, + "loss": 0.383, + "step": 33810 + }, + { + "epoch": 1.3058419243986255, + "grad_norm": 0.2249228060245514, + "learning_rate": 0.0001129464458087185, + "loss": 0.14, + "step": 33820 + }, + { + "epoch": 1.3062280396926522, + "grad_norm": 0.4820781946182251, + "learning_rate": 0.0001129207047891167, + "loss": 0.248, + "step": 33830 + }, + { + "epoch": 1.306614154986679, + "grad_norm": 2.2983391284942627, + "learning_rate": 0.00011289496376951491, + "loss": 0.2608, + "step": 33840 + }, + { + "epoch": 1.3070002702807058, + "grad_norm": 1.3315671682357788, + "learning_rate": 0.00011286922274991314, + "loss": 0.1727, + "step": 33850 + }, + { + "epoch": 1.3073863855747327, + "grad_norm": 2.060299873352051, + "learning_rate": 0.00011284348173031134, + "loss": 0.3097, + "step": 33860 + }, + { + "epoch": 1.3077725008687593, + "grad_norm": 2.096285581588745, + "learning_rate": 0.00011281774071070957, + "loss": 0.2305, + "step": 33870 + }, + { + "epoch": 1.3081586161627863, + "grad_norm": 0.4997636675834656, + "learning_rate": 0.00011279199969110777, + "loss": 0.1993, + "step": 33880 + }, + { + "epoch": 1.308544731456813, + "grad_norm": 0.58636474609375, + "learning_rate": 0.00011276625867150599, + "loss": 0.1873, + "step": 33890 + }, + { + "epoch": 1.3089308467508398, + "grad_norm": 0.9128592610359192, + "learning_rate": 0.00011274051765190419, + "loss": 0.1885, + "step": 33900 + }, + { + "epoch": 1.3093169620448666, + "grad_norm": 2.228043794631958, + "learning_rate": 0.0001127147766323024, + "loss": 0.3649, + "step": 33910 + }, + { + "epoch": 1.3097030773388934, + "grad_norm": 1.069002389907837, + "learning_rate": 0.00011268903561270063, + "loss": 0.5454, + "step": 33920 + }, + { + "epoch": 1.3100891926329201, + "grad_norm": 0.6207597851753235, + "learning_rate": 0.00011266329459309883, + "loss": 0.2329, + "step": 33930 + }, + { + "epoch": 1.3104753079269469, + "grad_norm": 1.262247920036316, + "learning_rate": 0.00011263755357349706, + "loss": 0.3437, + "step": 33940 + }, + { + "epoch": 1.3108614232209739, + "grad_norm": 1.7429994344711304, + "learning_rate": 0.00011261181255389526, + "loss": 0.228, + "step": 33950 + }, + { + "epoch": 1.3112475385150006, + "grad_norm": 0.646900475025177, + "learning_rate": 0.00011258607153429349, + "loss": 0.3739, + "step": 33960 + }, + { + "epoch": 1.3116336538090274, + "grad_norm": 1.8228782415390015, + "learning_rate": 0.0001125603305146917, + "loss": 0.2325, + "step": 33970 + }, + { + "epoch": 1.3120197691030542, + "grad_norm": 3.539228916168213, + "learning_rate": 0.0001125345894950899, + "loss": 0.194, + "step": 33980 + }, + { + "epoch": 1.312405884397081, + "grad_norm": 1.2801135778427124, + "learning_rate": 0.00011250884847548813, + "loss": 0.3069, + "step": 33990 + }, + { + "epoch": 1.3127919996911077, + "grad_norm": 3.6265695095062256, + "learning_rate": 0.00011248310745588633, + "loss": 0.3113, + "step": 34000 + }, + { + "epoch": 1.3131781149851345, + "grad_norm": 0.07370063662528992, + "learning_rate": 0.00011245736643628455, + "loss": 0.1449, + "step": 34010 + }, + { + "epoch": 1.3135642302791615, + "grad_norm": 1.0295637845993042, + "learning_rate": 0.00011243162541668275, + "loss": 0.22, + "step": 34020 + }, + { + "epoch": 1.313950345573188, + "grad_norm": 0.8803662061691284, + "learning_rate": 0.00011240588439708098, + "loss": 0.1368, + "step": 34030 + }, + { + "epoch": 1.314336460867215, + "grad_norm": 1.6597707271575928, + "learning_rate": 0.00011238014337747919, + "loss": 0.3038, + "step": 34040 + }, + { + "epoch": 1.3147225761612418, + "grad_norm": 2.115492343902588, + "learning_rate": 0.00011235440235787739, + "loss": 0.1754, + "step": 34050 + }, + { + "epoch": 1.3151086914552685, + "grad_norm": 0.8143919706344604, + "learning_rate": 0.00011232866133827562, + "loss": 0.3764, + "step": 34060 + }, + { + "epoch": 1.3154948067492953, + "grad_norm": 0.14369767904281616, + "learning_rate": 0.00011230292031867382, + "loss": 0.1742, + "step": 34070 + }, + { + "epoch": 1.315880922043322, + "grad_norm": 1.0129845142364502, + "learning_rate": 0.00011227717929907205, + "loss": 0.1458, + "step": 34080 + }, + { + "epoch": 1.316267037337349, + "grad_norm": 2.7300291061401367, + "learning_rate": 0.00011225143827947025, + "loss": 0.3939, + "step": 34090 + }, + { + "epoch": 1.3166531526313756, + "grad_norm": 0.20205609500408173, + "learning_rate": 0.00011222569725986847, + "loss": 0.29, + "step": 34100 + }, + { + "epoch": 1.3170392679254026, + "grad_norm": 1.8928464651107788, + "learning_rate": 0.00011219995624026669, + "loss": 0.1742, + "step": 34110 + }, + { + "epoch": 1.3174253832194294, + "grad_norm": 0.2639687955379486, + "learning_rate": 0.00011217421522066488, + "loss": 0.1745, + "step": 34120 + }, + { + "epoch": 1.3178114985134561, + "grad_norm": 0.5906389355659485, + "learning_rate": 0.00011214847420106311, + "loss": 0.2134, + "step": 34130 + }, + { + "epoch": 1.318197613807483, + "grad_norm": 0.9190629720687866, + "learning_rate": 0.00011212273318146131, + "loss": 0.2547, + "step": 34140 + }, + { + "epoch": 1.3185837291015097, + "grad_norm": 0.5749151110649109, + "learning_rate": 0.00011209699216185954, + "loss": 0.1688, + "step": 34150 + }, + { + "epoch": 1.3189698443955364, + "grad_norm": 0.82295823097229, + "learning_rate": 0.00011207125114225777, + "loss": 0.2884, + "step": 34160 + }, + { + "epoch": 1.3193559596895632, + "grad_norm": 0.07816460728645325, + "learning_rate": 0.00011204551012265597, + "loss": 0.2418, + "step": 34170 + }, + { + "epoch": 1.3197420749835902, + "grad_norm": 0.6417407393455505, + "learning_rate": 0.00011201976910305418, + "loss": 0.2557, + "step": 34180 + }, + { + "epoch": 1.320128190277617, + "grad_norm": 6.093267440795898, + "learning_rate": 0.00011199402808345238, + "loss": 0.3088, + "step": 34190 + }, + { + "epoch": 1.3205143055716437, + "grad_norm": 1.8861887454986572, + "learning_rate": 0.0001119682870638506, + "loss": 0.2204, + "step": 34200 + }, + { + "epoch": 1.3209004208656705, + "grad_norm": 2.3272714614868164, + "learning_rate": 0.0001119425460442488, + "loss": 0.2236, + "step": 34210 + }, + { + "epoch": 1.3212865361596973, + "grad_norm": 0.9608810544013977, + "learning_rate": 0.00011191680502464703, + "loss": 0.1897, + "step": 34220 + }, + { + "epoch": 1.321672651453724, + "grad_norm": 1.2157350778579712, + "learning_rate": 0.00011189106400504526, + "loss": 0.1526, + "step": 34230 + }, + { + "epoch": 1.3220587667477508, + "grad_norm": 1.6684671640396118, + "learning_rate": 0.00011186532298544346, + "loss": 0.3394, + "step": 34240 + }, + { + "epoch": 1.3224448820417778, + "grad_norm": 2.0432374477386475, + "learning_rate": 0.00011183958196584167, + "loss": 0.2183, + "step": 34250 + }, + { + "epoch": 1.3228309973358043, + "grad_norm": 0.9436892867088318, + "learning_rate": 0.00011181384094623988, + "loss": 0.2947, + "step": 34260 + }, + { + "epoch": 1.3232171126298313, + "grad_norm": 0.23260092735290527, + "learning_rate": 0.0001117880999266381, + "loss": 0.114, + "step": 34270 + }, + { + "epoch": 1.323603227923858, + "grad_norm": 1.2291594743728638, + "learning_rate": 0.0001117623589070363, + "loss": 0.3145, + "step": 34280 + }, + { + "epoch": 1.3239893432178849, + "grad_norm": 0.41411107778549194, + "learning_rate": 0.00011173661788743452, + "loss": 0.2937, + "step": 34290 + }, + { + "epoch": 1.3243754585119116, + "grad_norm": 2.354405164718628, + "learning_rate": 0.00011171087686783275, + "loss": 0.3933, + "step": 34300 + }, + { + "epoch": 1.3247615738059384, + "grad_norm": 2.6997978687286377, + "learning_rate": 0.00011168513584823095, + "loss": 0.1494, + "step": 34310 + }, + { + "epoch": 1.3251476890999654, + "grad_norm": 2.8430919647216797, + "learning_rate": 0.00011165939482862916, + "loss": 0.2869, + "step": 34320 + }, + { + "epoch": 1.325533804393992, + "grad_norm": 1.1737356185913086, + "learning_rate": 0.00011163365380902738, + "loss": 0.2792, + "step": 34330 + }, + { + "epoch": 1.325919919688019, + "grad_norm": 4.123973846435547, + "learning_rate": 0.00011160791278942559, + "loss": 0.5211, + "step": 34340 + }, + { + "epoch": 1.3263060349820457, + "grad_norm": 0.8862038850784302, + "learning_rate": 0.00011158217176982382, + "loss": 0.2976, + "step": 34350 + }, + { + "epoch": 1.3266921502760725, + "grad_norm": 1.8690590858459473, + "learning_rate": 0.00011155643075022202, + "loss": 0.2485, + "step": 34360 + }, + { + "epoch": 1.3270782655700992, + "grad_norm": 2.885589599609375, + "learning_rate": 0.00011153068973062024, + "loss": 0.3084, + "step": 34370 + }, + { + "epoch": 1.327464380864126, + "grad_norm": 0.9898788928985596, + "learning_rate": 0.00011150494871101844, + "loss": 0.261, + "step": 34380 + }, + { + "epoch": 1.3278504961581528, + "grad_norm": 0.6879653930664062, + "learning_rate": 0.00011147920769141667, + "loss": 0.2082, + "step": 34390 + }, + { + "epoch": 1.3282366114521795, + "grad_norm": 1.2619003057479858, + "learning_rate": 0.00011145346667181487, + "loss": 0.2402, + "step": 34400 + }, + { + "epoch": 1.3286227267462065, + "grad_norm": 1.1212007999420166, + "learning_rate": 0.00011142772565221308, + "loss": 0.3525, + "step": 34410 + }, + { + "epoch": 1.3290088420402333, + "grad_norm": 1.8431956768035889, + "learning_rate": 0.00011140198463261131, + "loss": 0.2212, + "step": 34420 + }, + { + "epoch": 1.32939495733426, + "grad_norm": 0.6185423731803894, + "learning_rate": 0.00011137624361300951, + "loss": 0.2455, + "step": 34430 + }, + { + "epoch": 1.3297810726282868, + "grad_norm": 2.3791301250457764, + "learning_rate": 0.00011135050259340774, + "loss": 0.1763, + "step": 34440 + }, + { + "epoch": 1.3301671879223136, + "grad_norm": 0.4928603768348694, + "learning_rate": 0.00011132476157380594, + "loss": 0.2381, + "step": 34450 + }, + { + "epoch": 1.3305533032163404, + "grad_norm": 1.5636029243469238, + "learning_rate": 0.00011129902055420416, + "loss": 0.1368, + "step": 34460 + }, + { + "epoch": 1.3309394185103671, + "grad_norm": 0.9425283074378967, + "learning_rate": 0.00011127327953460238, + "loss": 0.2825, + "step": 34470 + }, + { + "epoch": 1.3313255338043941, + "grad_norm": 1.2257115840911865, + "learning_rate": 0.00011124753851500058, + "loss": 0.2547, + "step": 34480 + }, + { + "epoch": 1.3317116490984207, + "grad_norm": 0.9416170716285706, + "learning_rate": 0.0001112217974953988, + "loss": 0.2766, + "step": 34490 + }, + { + "epoch": 1.3320977643924476, + "grad_norm": 0.5123847126960754, + "learning_rate": 0.000111196056475797, + "loss": 0.4733, + "step": 34500 + }, + { + "epoch": 1.3324838796864744, + "grad_norm": 1.5581384897232056, + "learning_rate": 0.00011117031545619523, + "loss": 0.1597, + "step": 34510 + }, + { + "epoch": 1.3328699949805012, + "grad_norm": 2.377333879470825, + "learning_rate": 0.00011114457443659343, + "loss": 0.209, + "step": 34520 + }, + { + "epoch": 1.333256110274528, + "grad_norm": 1.7840913534164429, + "learning_rate": 0.00011111883341699166, + "loss": 0.1759, + "step": 34530 + }, + { + "epoch": 1.3336422255685547, + "grad_norm": 1.1825993061065674, + "learning_rate": 0.00011109309239738987, + "loss": 0.2464, + "step": 34540 + }, + { + "epoch": 1.3340283408625815, + "grad_norm": 1.8859659433364868, + "learning_rate": 0.00011106735137778807, + "loss": 0.3539, + "step": 34550 + }, + { + "epoch": 1.3344144561566083, + "grad_norm": 1.9698175191879272, + "learning_rate": 0.0001110416103581863, + "loss": 0.3301, + "step": 34560 + }, + { + "epoch": 1.3348005714506352, + "grad_norm": 0.7649385333061218, + "learning_rate": 0.0001110158693385845, + "loss": 0.232, + "step": 34570 + }, + { + "epoch": 1.335186686744662, + "grad_norm": 0.56386399269104, + "learning_rate": 0.00011099012831898272, + "loss": 0.3425, + "step": 34580 + }, + { + "epoch": 1.3355728020386888, + "grad_norm": 2.956003189086914, + "learning_rate": 0.00011096438729938092, + "loss": 0.1518, + "step": 34590 + }, + { + "epoch": 1.3359589173327155, + "grad_norm": 2.612029552459717, + "learning_rate": 0.00011093864627977915, + "loss": 0.2765, + "step": 34600 + }, + { + "epoch": 1.3363450326267423, + "grad_norm": 0.9674397706985474, + "learning_rate": 0.00011091290526017736, + "loss": 0.303, + "step": 34610 + }, + { + "epoch": 1.336731147920769, + "grad_norm": 0.9578921794891357, + "learning_rate": 0.00011088716424057556, + "loss": 0.1405, + "step": 34620 + }, + { + "epoch": 1.3371172632147958, + "grad_norm": 2.168065071105957, + "learning_rate": 0.00011086142322097379, + "loss": 0.1914, + "step": 34630 + }, + { + "epoch": 1.3375033785088228, + "grad_norm": 1.3166526556015015, + "learning_rate": 0.00011083568220137199, + "loss": 0.4134, + "step": 34640 + }, + { + "epoch": 1.3378894938028496, + "grad_norm": 0.9082283973693848, + "learning_rate": 0.00011080994118177022, + "loss": 0.2693, + "step": 34650 + }, + { + "epoch": 1.3382756090968764, + "grad_norm": 2.203007698059082, + "learning_rate": 0.00011078420016216844, + "loss": 0.1847, + "step": 34660 + }, + { + "epoch": 1.3386617243909031, + "grad_norm": 0.8101674914360046, + "learning_rate": 0.00011075845914256664, + "loss": 0.3111, + "step": 34670 + }, + { + "epoch": 1.33904783968493, + "grad_norm": 1.9545695781707764, + "learning_rate": 0.00011073271812296486, + "loss": 0.3761, + "step": 34680 + }, + { + "epoch": 1.3394339549789567, + "grad_norm": 1.547581672668457, + "learning_rate": 0.00011070697710336306, + "loss": 0.2374, + "step": 34690 + }, + { + "epoch": 1.3398200702729834, + "grad_norm": 3.3519034385681152, + "learning_rate": 0.00011068123608376128, + "loss": 0.1957, + "step": 34700 + }, + { + "epoch": 1.3402061855670104, + "grad_norm": 1.5508599281311035, + "learning_rate": 0.00011065549506415948, + "loss": 0.4171, + "step": 34710 + }, + { + "epoch": 1.340592300861037, + "grad_norm": 1.8547546863555908, + "learning_rate": 0.00011062975404455771, + "loss": 0.1872, + "step": 34720 + }, + { + "epoch": 1.340978416155064, + "grad_norm": 1.4600756168365479, + "learning_rate": 0.00011060401302495594, + "loss": 0.3515, + "step": 34730 + }, + { + "epoch": 1.3413645314490907, + "grad_norm": 0.05774044618010521, + "learning_rate": 0.00011057827200535414, + "loss": 0.1604, + "step": 34740 + }, + { + "epoch": 1.3417506467431175, + "grad_norm": 2.8793342113494873, + "learning_rate": 0.00011055253098575235, + "loss": 0.3095, + "step": 34750 + }, + { + "epoch": 1.3421367620371443, + "grad_norm": 2.241042375564575, + "learning_rate": 0.00011052678996615055, + "loss": 0.2511, + "step": 34760 + }, + { + "epoch": 1.342522877331171, + "grad_norm": 1.9320632219314575, + "learning_rate": 0.00011050104894654878, + "loss": 0.4493, + "step": 34770 + }, + { + "epoch": 1.3429089926251978, + "grad_norm": 1.6483882665634155, + "learning_rate": 0.000110475307926947, + "loss": 0.217, + "step": 34780 + }, + { + "epoch": 1.3432951079192246, + "grad_norm": 0.9635765552520752, + "learning_rate": 0.0001104495669073452, + "loss": 0.5458, + "step": 34790 + }, + { + "epoch": 1.3436812232132516, + "grad_norm": 1.2436567544937134, + "learning_rate": 0.00011042382588774343, + "loss": 0.2857, + "step": 34800 + }, + { + "epoch": 1.3440673385072783, + "grad_norm": 2.8082425594329834, + "learning_rate": 0.00011039808486814163, + "loss": 0.3439, + "step": 34810 + }, + { + "epoch": 1.344453453801305, + "grad_norm": 1.0430901050567627, + "learning_rate": 0.00011037234384853984, + "loss": 0.1404, + "step": 34820 + }, + { + "epoch": 1.3448395690953319, + "grad_norm": 1.7387149333953857, + "learning_rate": 0.00011034660282893806, + "loss": 0.395, + "step": 34830 + }, + { + "epoch": 1.3452256843893586, + "grad_norm": 1.2713748216629028, + "learning_rate": 0.00011032086180933627, + "loss": 0.29, + "step": 34840 + }, + { + "epoch": 1.3456117996833854, + "grad_norm": 0.26068204641342163, + "learning_rate": 0.0001102951207897345, + "loss": 0.1814, + "step": 34850 + }, + { + "epoch": 1.3459979149774122, + "grad_norm": 2.4163243770599365, + "learning_rate": 0.0001102693797701327, + "loss": 0.2445, + "step": 34860 + }, + { + "epoch": 1.3463840302714392, + "grad_norm": 2.2439687252044678, + "learning_rate": 0.00011024363875053092, + "loss": 0.3338, + "step": 34870 + }, + { + "epoch": 1.346770145565466, + "grad_norm": 0.2822403609752655, + "learning_rate": 0.00011021789773092912, + "loss": 0.1648, + "step": 34880 + }, + { + "epoch": 1.3471562608594927, + "grad_norm": 0.07319017499685287, + "learning_rate": 0.00011019215671132734, + "loss": 0.107, + "step": 34890 + }, + { + "epoch": 1.3475423761535195, + "grad_norm": 0.9809044003486633, + "learning_rate": 0.00011016641569172555, + "loss": 0.256, + "step": 34900 + }, + { + "epoch": 1.3479284914475462, + "grad_norm": 0.5016226768493652, + "learning_rate": 0.00011014067467212376, + "loss": 0.3229, + "step": 34910 + }, + { + "epoch": 1.348314606741573, + "grad_norm": 1.3026005029678345, + "learning_rate": 0.00011011493365252199, + "loss": 0.2145, + "step": 34920 + }, + { + "epoch": 1.3487007220355998, + "grad_norm": 1.0752215385437012, + "learning_rate": 0.00011008919263292019, + "loss": 0.2355, + "step": 34930 + }, + { + "epoch": 1.3490868373296268, + "grad_norm": 2.2703003883361816, + "learning_rate": 0.00011006345161331842, + "loss": 0.2079, + "step": 34940 + }, + { + "epoch": 1.3494729526236533, + "grad_norm": 1.1323810815811157, + "learning_rate": 0.00011003771059371662, + "loss": 0.1015, + "step": 34950 + }, + { + "epoch": 1.3498590679176803, + "grad_norm": 0.10813555121421814, + "learning_rate": 0.00011001196957411484, + "loss": 0.4214, + "step": 34960 + }, + { + "epoch": 1.350245183211707, + "grad_norm": 0.07815568149089813, + "learning_rate": 0.00010998622855451306, + "loss": 0.1211, + "step": 34970 + }, + { + "epoch": 1.3506312985057338, + "grad_norm": 0.6748234629631042, + "learning_rate": 0.00010996048753491126, + "loss": 0.3508, + "step": 34980 + }, + { + "epoch": 1.3510174137997606, + "grad_norm": 1.8556997776031494, + "learning_rate": 0.00010993474651530948, + "loss": 0.2268, + "step": 34990 + }, + { + "epoch": 1.3514035290937874, + "grad_norm": 0.8696061372756958, + "learning_rate": 0.00010990900549570768, + "loss": 0.4321, + "step": 35000 + }, + { + "epoch": 1.3517896443878141, + "grad_norm": 0.42442765831947327, + "learning_rate": 0.00010988326447610591, + "loss": 0.1944, + "step": 35010 + }, + { + "epoch": 1.352175759681841, + "grad_norm": 1.0474554300308228, + "learning_rate": 0.00010985752345650411, + "loss": 0.1342, + "step": 35020 + }, + { + "epoch": 1.3525618749758679, + "grad_norm": 0.607037365436554, + "learning_rate": 0.00010983178243690234, + "loss": 0.2965, + "step": 35030 + }, + { + "epoch": 1.3529479902698947, + "grad_norm": 1.8160990476608276, + "learning_rate": 0.00010980604141730055, + "loss": 0.3192, + "step": 35040 + }, + { + "epoch": 1.3533341055639214, + "grad_norm": 2.0026509761810303, + "learning_rate": 0.00010978030039769875, + "loss": 0.3054, + "step": 35050 + }, + { + "epoch": 1.3537202208579482, + "grad_norm": 0.9203600883483887, + "learning_rate": 0.00010975455937809698, + "loss": 0.253, + "step": 35060 + }, + { + "epoch": 1.354106336151975, + "grad_norm": 0.33198195695877075, + "learning_rate": 0.00010972881835849518, + "loss": 0.3885, + "step": 35070 + }, + { + "epoch": 1.3544924514460017, + "grad_norm": 0.3201223611831665, + "learning_rate": 0.0001097030773388934, + "loss": 0.3029, + "step": 35080 + }, + { + "epoch": 1.3548785667400285, + "grad_norm": 1.2589943408966064, + "learning_rate": 0.0001096773363192916, + "loss": 0.4243, + "step": 35090 + }, + { + "epoch": 1.3552646820340555, + "grad_norm": 1.5106219053268433, + "learning_rate": 0.00010965159529968983, + "loss": 0.2585, + "step": 35100 + }, + { + "epoch": 1.3556507973280822, + "grad_norm": 1.429799199104309, + "learning_rate": 0.00010962585428008804, + "loss": 0.1961, + "step": 35110 + }, + { + "epoch": 1.356036912622109, + "grad_norm": 2.1211297512054443, + "learning_rate": 0.00010960011326048624, + "loss": 0.4057, + "step": 35120 + }, + { + "epoch": 1.3564230279161358, + "grad_norm": 2.5154731273651123, + "learning_rate": 0.00010957437224088447, + "loss": 0.3787, + "step": 35130 + }, + { + "epoch": 1.3568091432101625, + "grad_norm": 0.4914834201335907, + "learning_rate": 0.00010954863122128267, + "loss": 0.234, + "step": 35140 + }, + { + "epoch": 1.3571952585041893, + "grad_norm": 0.26685893535614014, + "learning_rate": 0.0001095228902016809, + "loss": 0.2841, + "step": 35150 + }, + { + "epoch": 1.357581373798216, + "grad_norm": 0.15462155640125275, + "learning_rate": 0.00010949714918207912, + "loss": 0.2269, + "step": 35160 + }, + { + "epoch": 1.357967489092243, + "grad_norm": 1.3887063264846802, + "learning_rate": 0.00010947140816247732, + "loss": 0.3455, + "step": 35170 + }, + { + "epoch": 1.3583536043862696, + "grad_norm": 0.786374032497406, + "learning_rate": 0.00010944566714287554, + "loss": 0.2897, + "step": 35180 + }, + { + "epoch": 1.3587397196802966, + "grad_norm": 1.100475549697876, + "learning_rate": 0.00010941992612327374, + "loss": 0.2892, + "step": 35190 + }, + { + "epoch": 1.3591258349743234, + "grad_norm": 0.7676102519035339, + "learning_rate": 0.00010939418510367196, + "loss": 0.1942, + "step": 35200 + }, + { + "epoch": 1.3595119502683501, + "grad_norm": 0.33462053537368774, + "learning_rate": 0.00010936844408407016, + "loss": 0.2872, + "step": 35210 + }, + { + "epoch": 1.359898065562377, + "grad_norm": 0.9294387698173523, + "learning_rate": 0.00010934270306446839, + "loss": 0.2617, + "step": 35220 + }, + { + "epoch": 1.3602841808564037, + "grad_norm": 0.3169979453086853, + "learning_rate": 0.00010931696204486662, + "loss": 0.2942, + "step": 35230 + }, + { + "epoch": 1.3606702961504304, + "grad_norm": 2.1339616775512695, + "learning_rate": 0.00010929122102526482, + "loss": 0.4448, + "step": 35240 + }, + { + "epoch": 1.3610564114444572, + "grad_norm": 0.9430062770843506, + "learning_rate": 0.00010926548000566303, + "loss": 0.2051, + "step": 35250 + }, + { + "epoch": 1.3614425267384842, + "grad_norm": 3.1187360286712646, + "learning_rate": 0.00010923973898606123, + "loss": 0.2274, + "step": 35260 + }, + { + "epoch": 1.361828642032511, + "grad_norm": 1.4727579355239868, + "learning_rate": 0.00010921399796645946, + "loss": 0.3757, + "step": 35270 + }, + { + "epoch": 1.3622147573265377, + "grad_norm": 2.157560348510742, + "learning_rate": 0.00010918825694685768, + "loss": 0.3096, + "step": 35280 + }, + { + "epoch": 1.3626008726205645, + "grad_norm": 0.33457377552986145, + "learning_rate": 0.00010916251592725588, + "loss": 0.1489, + "step": 35290 + }, + { + "epoch": 1.3629869879145913, + "grad_norm": 0.9005904197692871, + "learning_rate": 0.00010913677490765411, + "loss": 0.1826, + "step": 35300 + }, + { + "epoch": 1.363373103208618, + "grad_norm": 2.1222829818725586, + "learning_rate": 0.00010911103388805231, + "loss": 0.1965, + "step": 35310 + }, + { + "epoch": 1.3637592185026448, + "grad_norm": 1.3881357908248901, + "learning_rate": 0.00010908529286845052, + "loss": 0.1791, + "step": 35320 + }, + { + "epoch": 1.3641453337966718, + "grad_norm": 1.7574503421783447, + "learning_rate": 0.00010905955184884872, + "loss": 0.3316, + "step": 35330 + }, + { + "epoch": 1.3645314490906983, + "grad_norm": 0.1967727392911911, + "learning_rate": 0.00010903381082924695, + "loss": 0.2331, + "step": 35340 + }, + { + "epoch": 1.3649175643847253, + "grad_norm": 0.8974360823631287, + "learning_rate": 0.00010900806980964518, + "loss": 0.2589, + "step": 35350 + }, + { + "epoch": 1.365303679678752, + "grad_norm": 2.0996744632720947, + "learning_rate": 0.00010898232879004338, + "loss": 0.3663, + "step": 35360 + }, + { + "epoch": 1.3656897949727789, + "grad_norm": 0.5678316354751587, + "learning_rate": 0.0001089565877704416, + "loss": 0.1729, + "step": 35370 + }, + { + "epoch": 1.3660759102668056, + "grad_norm": 2.3381874561309814, + "learning_rate": 0.0001089308467508398, + "loss": 0.1615, + "step": 35380 + }, + { + "epoch": 1.3664620255608324, + "grad_norm": 1.0276836156845093, + "learning_rate": 0.00010890510573123802, + "loss": 0.3359, + "step": 35390 + }, + { + "epoch": 1.3668481408548594, + "grad_norm": 2.4374940395355225, + "learning_rate": 0.00010887936471163622, + "loss": 0.2435, + "step": 35400 + }, + { + "epoch": 1.367234256148886, + "grad_norm": 0.45221665501594543, + "learning_rate": 0.00010885362369203444, + "loss": 0.2555, + "step": 35410 + }, + { + "epoch": 1.367620371442913, + "grad_norm": 2.608090400695801, + "learning_rate": 0.00010882788267243267, + "loss": 0.2465, + "step": 35420 + }, + { + "epoch": 1.3680064867369397, + "grad_norm": 1.4186642169952393, + "learning_rate": 0.00010880214165283087, + "loss": 0.1674, + "step": 35430 + }, + { + "epoch": 1.3683926020309665, + "grad_norm": 0.659479022026062, + "learning_rate": 0.0001087764006332291, + "loss": 0.2926, + "step": 35440 + }, + { + "epoch": 1.3687787173249932, + "grad_norm": 0.9219567179679871, + "learning_rate": 0.0001087506596136273, + "loss": 0.2001, + "step": 35450 + }, + { + "epoch": 1.36916483261902, + "grad_norm": 0.8070804476737976, + "learning_rate": 0.00010872491859402551, + "loss": 0.2178, + "step": 35460 + }, + { + "epoch": 1.3695509479130468, + "grad_norm": 2.9981069564819336, + "learning_rate": 0.00010869917757442374, + "loss": 0.3079, + "step": 35470 + }, + { + "epoch": 1.3699370632070735, + "grad_norm": 0.7891242504119873, + "learning_rate": 0.00010867343655482194, + "loss": 0.2765, + "step": 35480 + }, + { + "epoch": 1.3703231785011005, + "grad_norm": 1.448637843132019, + "learning_rate": 0.00010864769553522016, + "loss": 0.3521, + "step": 35490 + }, + { + "epoch": 1.3707092937951273, + "grad_norm": 0.07628043740987778, + "learning_rate": 0.00010862195451561836, + "loss": 0.2083, + "step": 35500 + }, + { + "epoch": 1.371095409089154, + "grad_norm": 0.7549735307693481, + "learning_rate": 0.00010859621349601659, + "loss": 0.2536, + "step": 35510 + }, + { + "epoch": 1.3714815243831808, + "grad_norm": 1.3548041582107544, + "learning_rate": 0.00010857047247641479, + "loss": 0.251, + "step": 35520 + }, + { + "epoch": 1.3718676396772076, + "grad_norm": 0.530010998249054, + "learning_rate": 0.000108544731456813, + "loss": 0.1917, + "step": 35530 + }, + { + "epoch": 1.3722537549712344, + "grad_norm": 0.4148992896080017, + "learning_rate": 0.00010851899043721123, + "loss": 0.335, + "step": 35540 + }, + { + "epoch": 1.3726398702652611, + "grad_norm": 1.5118776559829712, + "learning_rate": 0.00010849324941760943, + "loss": 0.2159, + "step": 35550 + }, + { + "epoch": 1.3730259855592881, + "grad_norm": 1.036889910697937, + "learning_rate": 0.00010846750839800766, + "loss": 0.2975, + "step": 35560 + }, + { + "epoch": 1.3734121008533147, + "grad_norm": 1.724263072013855, + "learning_rate": 0.00010844176737840586, + "loss": 0.1476, + "step": 35570 + }, + { + "epoch": 1.3737982161473417, + "grad_norm": 1.599007487297058, + "learning_rate": 0.00010841602635880408, + "loss": 0.2539, + "step": 35580 + }, + { + "epoch": 1.3741843314413684, + "grad_norm": 2.9119279384613037, + "learning_rate": 0.00010839028533920228, + "loss": 0.2688, + "step": 35590 + }, + { + "epoch": 1.3745704467353952, + "grad_norm": 1.8647874593734741, + "learning_rate": 0.00010836454431960051, + "loss": 0.4158, + "step": 35600 + }, + { + "epoch": 1.374956562029422, + "grad_norm": 3.925290822982788, + "learning_rate": 0.00010833880329999872, + "loss": 0.3333, + "step": 35610 + }, + { + "epoch": 1.3753426773234487, + "grad_norm": 0.7124634385108948, + "learning_rate": 0.00010831306228039692, + "loss": 0.1069, + "step": 35620 + }, + { + "epoch": 1.3757287926174757, + "grad_norm": 1.303579330444336, + "learning_rate": 0.00010828732126079515, + "loss": 0.2898, + "step": 35630 + }, + { + "epoch": 1.3761149079115023, + "grad_norm": 3.921804189682007, + "learning_rate": 0.00010826158024119335, + "loss": 0.4212, + "step": 35640 + }, + { + "epoch": 1.3765010232055293, + "grad_norm": 1.3194564580917358, + "learning_rate": 0.00010823583922159158, + "loss": 0.2771, + "step": 35650 + }, + { + "epoch": 1.376887138499556, + "grad_norm": 1.4237637519836426, + "learning_rate": 0.00010821009820198979, + "loss": 0.2463, + "step": 35660 + }, + { + "epoch": 1.3772732537935828, + "grad_norm": 1.8165888786315918, + "learning_rate": 0.000108184357182388, + "loss": 0.291, + "step": 35670 + }, + { + "epoch": 1.3776593690876096, + "grad_norm": 1.1056426763534546, + "learning_rate": 0.00010815861616278622, + "loss": 0.2525, + "step": 35680 + }, + { + "epoch": 1.3780454843816363, + "grad_norm": 1.483189582824707, + "learning_rate": 0.00010813287514318442, + "loss": 0.1569, + "step": 35690 + }, + { + "epoch": 1.378431599675663, + "grad_norm": 1.0666841268539429, + "learning_rate": 0.00010810713412358264, + "loss": 0.235, + "step": 35700 + }, + { + "epoch": 1.3788177149696899, + "grad_norm": 1.0299845933914185, + "learning_rate": 0.00010808139310398084, + "loss": 0.3892, + "step": 35710 + }, + { + "epoch": 1.3792038302637168, + "grad_norm": 2.3474409580230713, + "learning_rate": 0.00010805565208437907, + "loss": 0.3417, + "step": 35720 + }, + { + "epoch": 1.3795899455577436, + "grad_norm": 1.7456315755844116, + "learning_rate": 0.0001080299110647773, + "loss": 0.2538, + "step": 35730 + }, + { + "epoch": 1.3799760608517704, + "grad_norm": 2.866103410720825, + "learning_rate": 0.0001080041700451755, + "loss": 0.1619, + "step": 35740 + }, + { + "epoch": 1.3803621761457971, + "grad_norm": 0.29136407375335693, + "learning_rate": 0.00010797842902557371, + "loss": 0.2692, + "step": 35750 + }, + { + "epoch": 1.380748291439824, + "grad_norm": 0.8046161532402039, + "learning_rate": 0.00010795268800597191, + "loss": 0.1575, + "step": 35760 + }, + { + "epoch": 1.3811344067338507, + "grad_norm": 0.6451787352561951, + "learning_rate": 0.00010792694698637014, + "loss": 0.4914, + "step": 35770 + }, + { + "epoch": 1.3815205220278775, + "grad_norm": 0.7289161086082458, + "learning_rate": 0.00010790120596676836, + "loss": 0.1895, + "step": 35780 + }, + { + "epoch": 1.3819066373219044, + "grad_norm": 0.8300430178642273, + "learning_rate": 0.00010787546494716656, + "loss": 0.3663, + "step": 35790 + }, + { + "epoch": 1.382292752615931, + "grad_norm": 0.17713364958763123, + "learning_rate": 0.00010784972392756479, + "loss": 0.3189, + "step": 35800 + }, + { + "epoch": 1.382678867909958, + "grad_norm": 0.903222918510437, + "learning_rate": 0.00010782398290796299, + "loss": 0.1577, + "step": 35810 + }, + { + "epoch": 1.3830649832039847, + "grad_norm": 0.08617932349443436, + "learning_rate": 0.0001077982418883612, + "loss": 0.2872, + "step": 35820 + }, + { + "epoch": 1.3834510984980115, + "grad_norm": 1.9590895175933838, + "learning_rate": 0.0001077725008687594, + "loss": 0.2907, + "step": 35830 + }, + { + "epoch": 1.3838372137920383, + "grad_norm": 1.2515161037445068, + "learning_rate": 0.00010774675984915763, + "loss": 0.177, + "step": 35840 + }, + { + "epoch": 1.384223329086065, + "grad_norm": 1.6171292066574097, + "learning_rate": 0.00010772101882955586, + "loss": 0.2321, + "step": 35850 + }, + { + "epoch": 1.3846094443800918, + "grad_norm": 0.13681405782699585, + "learning_rate": 0.00010769527780995406, + "loss": 0.247, + "step": 35860 + }, + { + "epoch": 1.3849955596741186, + "grad_norm": 1.1949968338012695, + "learning_rate": 0.00010766953679035228, + "loss": 0.2978, + "step": 35870 + }, + { + "epoch": 1.3853816749681456, + "grad_norm": 0.17001692950725555, + "learning_rate": 0.00010764379577075048, + "loss": 0.3241, + "step": 35880 + }, + { + "epoch": 1.3857677902621723, + "grad_norm": 0.8227952122688293, + "learning_rate": 0.0001076180547511487, + "loss": 0.3499, + "step": 35890 + }, + { + "epoch": 1.386153905556199, + "grad_norm": 1.4185482263565063, + "learning_rate": 0.0001075923137315469, + "loss": 0.3109, + "step": 35900 + }, + { + "epoch": 1.3865400208502259, + "grad_norm": 0.9533351063728333, + "learning_rate": 0.00010756657271194512, + "loss": 0.24, + "step": 35910 + }, + { + "epoch": 1.3869261361442526, + "grad_norm": 1.077789306640625, + "learning_rate": 0.00010754083169234335, + "loss": 0.2662, + "step": 35920 + }, + { + "epoch": 1.3873122514382794, + "grad_norm": 1.3528363704681396, + "learning_rate": 0.00010751509067274155, + "loss": 0.1623, + "step": 35930 + }, + { + "epoch": 1.3876983667323062, + "grad_norm": 0.25122806429862976, + "learning_rate": 0.00010748934965313978, + "loss": 0.1639, + "step": 35940 + }, + { + "epoch": 1.3880844820263332, + "grad_norm": 0.9446159601211548, + "learning_rate": 0.00010746360863353798, + "loss": 0.2035, + "step": 35950 + }, + { + "epoch": 1.38847059732036, + "grad_norm": 1.2258719205856323, + "learning_rate": 0.00010743786761393619, + "loss": 0.4247, + "step": 35960 + }, + { + "epoch": 1.3888567126143867, + "grad_norm": 3.117729663848877, + "learning_rate": 0.00010741212659433442, + "loss": 0.2569, + "step": 35970 + }, + { + "epoch": 1.3892428279084135, + "grad_norm": 0.6333123445510864, + "learning_rate": 0.00010738638557473262, + "loss": 0.1843, + "step": 35980 + }, + { + "epoch": 1.3896289432024402, + "grad_norm": 1.49360191822052, + "learning_rate": 0.00010736064455513084, + "loss": 0.2318, + "step": 35990 + }, + { + "epoch": 1.390015058496467, + "grad_norm": 3.9082753658294678, + "learning_rate": 0.00010733490353552904, + "loss": 0.2037, + "step": 36000 + }, + { + "epoch": 1.3904011737904938, + "grad_norm": 0.5687323808670044, + "learning_rate": 0.00010730916251592727, + "loss": 0.0871, + "step": 36010 + }, + { + "epoch": 1.3907872890845208, + "grad_norm": 0.3480868339538574, + "learning_rate": 0.00010728342149632547, + "loss": 0.4009, + "step": 36020 + }, + { + "epoch": 1.3911734043785473, + "grad_norm": 1.325042486190796, + "learning_rate": 0.00010725768047672368, + "loss": 0.5143, + "step": 36030 + }, + { + "epoch": 1.3915595196725743, + "grad_norm": 2.114786386489868, + "learning_rate": 0.00010723193945712191, + "loss": 0.2944, + "step": 36040 + }, + { + "epoch": 1.391945634966601, + "grad_norm": 1.716272234916687, + "learning_rate": 0.00010720619843752011, + "loss": 0.3793, + "step": 36050 + }, + { + "epoch": 1.3923317502606278, + "grad_norm": 2.057535171508789, + "learning_rate": 0.00010718045741791834, + "loss": 0.199, + "step": 36060 + }, + { + "epoch": 1.3927178655546546, + "grad_norm": 2.371248483657837, + "learning_rate": 0.00010715471639831653, + "loss": 0.3276, + "step": 36070 + }, + { + "epoch": 1.3931039808486814, + "grad_norm": 0.1784186065196991, + "learning_rate": 0.00010712897537871476, + "loss": 0.0922, + "step": 36080 + }, + { + "epoch": 1.3934900961427081, + "grad_norm": 0.7923040390014648, + "learning_rate": 0.00010710323435911298, + "loss": 0.1599, + "step": 36090 + }, + { + "epoch": 1.393876211436735, + "grad_norm": 1.9013831615447998, + "learning_rate": 0.00010707749333951117, + "loss": 0.4997, + "step": 36100 + }, + { + "epoch": 1.394262326730762, + "grad_norm": 1.8659415245056152, + "learning_rate": 0.0001070517523199094, + "loss": 0.1851, + "step": 36110 + }, + { + "epoch": 1.3946484420247887, + "grad_norm": 2.5775375366210938, + "learning_rate": 0.0001070260113003076, + "loss": 0.2384, + "step": 36120 + }, + { + "epoch": 1.3950345573188154, + "grad_norm": 0.21943879127502441, + "learning_rate": 0.00010700027028070583, + "loss": 0.3996, + "step": 36130 + }, + { + "epoch": 1.3954206726128422, + "grad_norm": 1.1734743118286133, + "learning_rate": 0.00010697452926110403, + "loss": 0.1737, + "step": 36140 + }, + { + "epoch": 1.395806787906869, + "grad_norm": 0.697695791721344, + "learning_rate": 0.00010694878824150225, + "loss": 0.2778, + "step": 36150 + }, + { + "epoch": 1.3961929032008957, + "grad_norm": 3.2881579399108887, + "learning_rate": 0.00010692304722190047, + "loss": 0.3198, + "step": 36160 + }, + { + "epoch": 1.3965790184949225, + "grad_norm": 0.1592467725276947, + "learning_rate": 0.00010689730620229868, + "loss": 0.2591, + "step": 36170 + }, + { + "epoch": 1.3969651337889495, + "grad_norm": 0.4579029083251953, + "learning_rate": 0.0001068715651826969, + "loss": 0.2984, + "step": 36180 + }, + { + "epoch": 1.3973512490829763, + "grad_norm": 0.4286015033721924, + "learning_rate": 0.0001068458241630951, + "loss": 0.4098, + "step": 36190 + }, + { + "epoch": 1.397737364377003, + "grad_norm": 1.7824127674102783, + "learning_rate": 0.00010682008314349332, + "loss": 0.2446, + "step": 36200 + }, + { + "epoch": 1.3981234796710298, + "grad_norm": 0.8584449887275696, + "learning_rate": 0.00010679434212389152, + "loss": 0.1415, + "step": 36210 + }, + { + "epoch": 1.3985095949650566, + "grad_norm": 1.1699339151382446, + "learning_rate": 0.00010676860110428975, + "loss": 0.1706, + "step": 36220 + }, + { + "epoch": 1.3988957102590833, + "grad_norm": 2.615877389907837, + "learning_rate": 0.00010674286008468796, + "loss": 0.2605, + "step": 36230 + }, + { + "epoch": 1.39928182555311, + "grad_norm": 2.182037591934204, + "learning_rate": 0.00010671711906508617, + "loss": 0.2067, + "step": 36240 + }, + { + "epoch": 1.399667940847137, + "grad_norm": 2.183263063430786, + "learning_rate": 0.00010669137804548439, + "loss": 0.4704, + "step": 36250 + }, + { + "epoch": 1.4000540561411636, + "grad_norm": 3.505791187286377, + "learning_rate": 0.00010666563702588259, + "loss": 0.3385, + "step": 36260 + }, + { + "epoch": 1.4004401714351906, + "grad_norm": 1.2262030839920044, + "learning_rate": 0.00010663989600628081, + "loss": 0.3999, + "step": 36270 + }, + { + "epoch": 1.4008262867292174, + "grad_norm": 2.4024577140808105, + "learning_rate": 0.00010661415498667904, + "loss": 0.1984, + "step": 36280 + }, + { + "epoch": 1.4012124020232442, + "grad_norm": 0.4166090786457062, + "learning_rate": 0.00010658841396707724, + "loss": 0.2448, + "step": 36290 + }, + { + "epoch": 1.401598517317271, + "grad_norm": 0.422590047121048, + "learning_rate": 0.00010656267294747547, + "loss": 0.2464, + "step": 36300 + }, + { + "epoch": 1.4019846326112977, + "grad_norm": 2.287503480911255, + "learning_rate": 0.00010653693192787367, + "loss": 0.1621, + "step": 36310 + }, + { + "epoch": 1.4023707479053245, + "grad_norm": 0.8126110434532166, + "learning_rate": 0.00010651119090827188, + "loss": 0.2097, + "step": 36320 + }, + { + "epoch": 1.4027568631993512, + "grad_norm": 0.683016836643219, + "learning_rate": 0.00010648544988867008, + "loss": 0.1512, + "step": 36330 + }, + { + "epoch": 1.4031429784933782, + "grad_norm": 1.5477893352508545, + "learning_rate": 0.00010645970886906831, + "loss": 0.2169, + "step": 36340 + }, + { + "epoch": 1.403529093787405, + "grad_norm": 2.183166265487671, + "learning_rate": 0.00010643396784946653, + "loss": 0.4307, + "step": 36350 + }, + { + "epoch": 1.4039152090814317, + "grad_norm": 1.5782747268676758, + "learning_rate": 0.00010640822682986473, + "loss": 0.291, + "step": 36360 + }, + { + "epoch": 1.4043013243754585, + "grad_norm": 1.2823392152786255, + "learning_rate": 0.00010638248581026296, + "loss": 0.315, + "step": 36370 + }, + { + "epoch": 1.4046874396694853, + "grad_norm": 2.0126500129699707, + "learning_rate": 0.00010635674479066116, + "loss": 0.2565, + "step": 36380 + }, + { + "epoch": 1.405073554963512, + "grad_norm": 2.490217447280884, + "learning_rate": 0.00010633100377105937, + "loss": 0.2665, + "step": 36390 + }, + { + "epoch": 1.4054596702575388, + "grad_norm": 1.2054855823516846, + "learning_rate": 0.00010630526275145757, + "loss": 0.5455, + "step": 36400 + }, + { + "epoch": 1.4058457855515658, + "grad_norm": 0.7968757748603821, + "learning_rate": 0.0001062795217318558, + "loss": 0.333, + "step": 36410 + }, + { + "epoch": 1.4062319008455926, + "grad_norm": 1.1027718782424927, + "learning_rate": 0.00010625378071225403, + "loss": 0.2019, + "step": 36420 + }, + { + "epoch": 1.4066180161396193, + "grad_norm": 1.987302541732788, + "learning_rate": 0.00010622803969265223, + "loss": 0.3159, + "step": 36430 + }, + { + "epoch": 1.407004131433646, + "grad_norm": 1.5426512956619263, + "learning_rate": 0.00010620229867305045, + "loss": 0.3759, + "step": 36440 + }, + { + "epoch": 1.4073902467276729, + "grad_norm": 0.5228156447410583, + "learning_rate": 0.00010617655765344865, + "loss": 0.2465, + "step": 36450 + }, + { + "epoch": 1.4077763620216996, + "grad_norm": 0.46890121698379517, + "learning_rate": 0.00010615081663384687, + "loss": 0.302, + "step": 36460 + }, + { + "epoch": 1.4081624773157264, + "grad_norm": 2.1506495475769043, + "learning_rate": 0.0001061250756142451, + "loss": 0.2569, + "step": 36470 + }, + { + "epoch": 1.4085485926097534, + "grad_norm": 2.307468891143799, + "learning_rate": 0.0001060993345946433, + "loss": 0.2009, + "step": 36480 + }, + { + "epoch": 1.40893470790378, + "grad_norm": 0.07033026963472366, + "learning_rate": 0.00010607359357504152, + "loss": 0.1728, + "step": 36490 + }, + { + "epoch": 1.409320823197807, + "grad_norm": 0.3262972831726074, + "learning_rate": 0.00010604785255543972, + "loss": 0.2905, + "step": 36500 + }, + { + "epoch": 1.4097069384918337, + "grad_norm": 0.755646824836731, + "learning_rate": 0.00010602211153583795, + "loss": 0.3287, + "step": 36510 + }, + { + "epoch": 1.4100930537858605, + "grad_norm": 1.1180161237716675, + "learning_rate": 0.00010599637051623615, + "loss": 0.2743, + "step": 36520 + }, + { + "epoch": 1.4104791690798872, + "grad_norm": 1.2358392477035522, + "learning_rate": 0.00010597062949663436, + "loss": 0.2672, + "step": 36530 + }, + { + "epoch": 1.410865284373914, + "grad_norm": 1.4188988208770752, + "learning_rate": 0.00010594488847703259, + "loss": 0.2552, + "step": 36540 + }, + { + "epoch": 1.4112513996679408, + "grad_norm": 1.4727978706359863, + "learning_rate": 0.00010591914745743079, + "loss": 0.2256, + "step": 36550 + }, + { + "epoch": 1.4116375149619675, + "grad_norm": 0.08973213285207748, + "learning_rate": 0.00010589340643782901, + "loss": 0.224, + "step": 36560 + }, + { + "epoch": 1.4120236302559945, + "grad_norm": 0.9915102124214172, + "learning_rate": 0.00010586766541822721, + "loss": 0.195, + "step": 36570 + }, + { + "epoch": 1.4124097455500213, + "grad_norm": 0.8524800539016724, + "learning_rate": 0.00010584192439862544, + "loss": 0.1492, + "step": 36580 + }, + { + "epoch": 1.412795860844048, + "grad_norm": 1.8414466381072998, + "learning_rate": 0.00010581618337902365, + "loss": 0.3412, + "step": 36590 + }, + { + "epoch": 1.4131819761380748, + "grad_norm": 2.596547842025757, + "learning_rate": 0.00010579044235942185, + "loss": 0.3469, + "step": 36600 + }, + { + "epoch": 1.4135680914321016, + "grad_norm": 0.5367813110351562, + "learning_rate": 0.00010576470133982008, + "loss": 0.2274, + "step": 36610 + }, + { + "epoch": 1.4139542067261284, + "grad_norm": 2.70858097076416, + "learning_rate": 0.00010573896032021828, + "loss": 0.171, + "step": 36620 + }, + { + "epoch": 1.4143403220201551, + "grad_norm": 1.7077667713165283, + "learning_rate": 0.00010571321930061651, + "loss": 0.2638, + "step": 36630 + }, + { + "epoch": 1.4147264373141821, + "grad_norm": 0.9189953804016113, + "learning_rate": 0.00010568747828101471, + "loss": 0.2283, + "step": 36640 + }, + { + "epoch": 1.4151125526082087, + "grad_norm": 2.0240087509155273, + "learning_rate": 0.00010566173726141293, + "loss": 0.3166, + "step": 36650 + }, + { + "epoch": 1.4154986679022357, + "grad_norm": 3.3304507732391357, + "learning_rate": 0.00010563599624181115, + "loss": 0.3046, + "step": 36660 + }, + { + "epoch": 1.4158847831962624, + "grad_norm": 1.955929160118103, + "learning_rate": 0.00010561025522220935, + "loss": 0.2653, + "step": 36670 + }, + { + "epoch": 1.4162708984902892, + "grad_norm": 2.2465381622314453, + "learning_rate": 0.00010558451420260757, + "loss": 0.3119, + "step": 36680 + }, + { + "epoch": 1.416657013784316, + "grad_norm": 2.108614921569824, + "learning_rate": 0.00010555877318300577, + "loss": 0.2872, + "step": 36690 + }, + { + "epoch": 1.4170431290783427, + "grad_norm": 1.35493004322052, + "learning_rate": 0.000105533032163404, + "loss": 0.1491, + "step": 36700 + }, + { + "epoch": 1.4174292443723697, + "grad_norm": 0.9102393984794617, + "learning_rate": 0.0001055072911438022, + "loss": 0.1565, + "step": 36710 + }, + { + "epoch": 1.4178153596663963, + "grad_norm": 1.7878345251083374, + "learning_rate": 0.00010548155012420043, + "loss": 0.4864, + "step": 36720 + }, + { + "epoch": 1.4182014749604233, + "grad_norm": 1.4333637952804565, + "learning_rate": 0.00010545580910459864, + "loss": 0.3246, + "step": 36730 + }, + { + "epoch": 1.41858759025445, + "grad_norm": 0.0966360941529274, + "learning_rate": 0.00010543006808499684, + "loss": 0.1758, + "step": 36740 + }, + { + "epoch": 1.4189737055484768, + "grad_norm": 0.30043545365333557, + "learning_rate": 0.00010540432706539507, + "loss": 0.1811, + "step": 36750 + }, + { + "epoch": 1.4193598208425036, + "grad_norm": 1.7705951929092407, + "learning_rate": 0.00010537858604579327, + "loss": 0.1824, + "step": 36760 + }, + { + "epoch": 1.4197459361365303, + "grad_norm": 1.2025195360183716, + "learning_rate": 0.0001053528450261915, + "loss": 0.1463, + "step": 36770 + }, + { + "epoch": 1.420132051430557, + "grad_norm": 0.3154304325580597, + "learning_rate": 0.00010532710400658972, + "loss": 0.1817, + "step": 36780 + }, + { + "epoch": 1.4205181667245839, + "grad_norm": 3.392331838607788, + "learning_rate": 0.00010530136298698792, + "loss": 0.2938, + "step": 36790 + }, + { + "epoch": 1.4209042820186109, + "grad_norm": 0.7256132364273071, + "learning_rate": 0.00010527562196738613, + "loss": 0.28, + "step": 36800 + }, + { + "epoch": 1.4212903973126376, + "grad_norm": 1.9007991552352905, + "learning_rate": 0.00010524988094778435, + "loss": 0.2306, + "step": 36810 + }, + { + "epoch": 1.4216765126066644, + "grad_norm": 3.21189546585083, + "learning_rate": 0.00010522413992818256, + "loss": 0.2945, + "step": 36820 + }, + { + "epoch": 1.4220626279006912, + "grad_norm": 0.20476600527763367, + "learning_rate": 0.00010519839890858076, + "loss": 0.3129, + "step": 36830 + }, + { + "epoch": 1.422448743194718, + "grad_norm": 1.4225107431411743, + "learning_rate": 0.00010517265788897899, + "loss": 0.2262, + "step": 36840 + }, + { + "epoch": 1.4228348584887447, + "grad_norm": 1.203728437423706, + "learning_rate": 0.00010514691686937721, + "loss": 0.176, + "step": 36850 + }, + { + "epoch": 1.4232209737827715, + "grad_norm": 0.3001759648323059, + "learning_rate": 0.00010512117584977541, + "loss": 0.218, + "step": 36860 + }, + { + "epoch": 1.4236070890767984, + "grad_norm": 2.3143389225006104, + "learning_rate": 0.00010509543483017363, + "loss": 0.1562, + "step": 36870 + }, + { + "epoch": 1.423993204370825, + "grad_norm": 0.5087364912033081, + "learning_rate": 0.00010506969381057184, + "loss": 0.1207, + "step": 36880 + }, + { + "epoch": 1.424379319664852, + "grad_norm": 1.6521960496902466, + "learning_rate": 0.00010504395279097005, + "loss": 0.4756, + "step": 36890 + }, + { + "epoch": 1.4247654349588788, + "grad_norm": 0.09236706793308258, + "learning_rate": 0.00010501821177136825, + "loss": 0.1755, + "step": 36900 + }, + { + "epoch": 1.4251515502529055, + "grad_norm": 0.2143094390630722, + "learning_rate": 0.00010499247075176648, + "loss": 0.2126, + "step": 36910 + }, + { + "epoch": 1.4255376655469323, + "grad_norm": 1.210170865058899, + "learning_rate": 0.00010496672973216471, + "loss": 0.1391, + "step": 36920 + }, + { + "epoch": 1.425923780840959, + "grad_norm": 0.6248244047164917, + "learning_rate": 0.00010494098871256291, + "loss": 0.2898, + "step": 36930 + }, + { + "epoch": 1.426309896134986, + "grad_norm": 2.1100337505340576, + "learning_rate": 0.00010491524769296113, + "loss": 0.1198, + "step": 36940 + }, + { + "epoch": 1.4266960114290126, + "grad_norm": 2.5673348903656006, + "learning_rate": 0.00010488950667335933, + "loss": 0.2284, + "step": 36950 + }, + { + "epoch": 1.4270821267230396, + "grad_norm": 1.868195652961731, + "learning_rate": 0.00010486376565375755, + "loss": 0.3738, + "step": 36960 + }, + { + "epoch": 1.4274682420170663, + "grad_norm": 1.0951671600341797, + "learning_rate": 0.00010483802463415577, + "loss": 0.1424, + "step": 36970 + }, + { + "epoch": 1.4278543573110931, + "grad_norm": 0.4791143536567688, + "learning_rate": 0.00010481228361455397, + "loss": 0.1445, + "step": 36980 + }, + { + "epoch": 1.4282404726051199, + "grad_norm": 3.2477540969848633, + "learning_rate": 0.0001047865425949522, + "loss": 0.2676, + "step": 36990 + }, + { + "epoch": 1.4286265878991466, + "grad_norm": 0.8082342147827148, + "learning_rate": 0.0001047608015753504, + "loss": 0.2295, + "step": 37000 + }, + { + "epoch": 1.4290127031931734, + "grad_norm": 2.3048954010009766, + "learning_rate": 0.00010473506055574863, + "loss": 0.1996, + "step": 37010 + }, + { + "epoch": 1.4293988184872002, + "grad_norm": 0.42648783326148987, + "learning_rate": 0.00010470931953614683, + "loss": 0.2549, + "step": 37020 + }, + { + "epoch": 1.4297849337812272, + "grad_norm": 1.8794362545013428, + "learning_rate": 0.00010468357851654504, + "loss": 0.1929, + "step": 37030 + }, + { + "epoch": 1.430171049075254, + "grad_norm": 0.9679039120674133, + "learning_rate": 0.00010465783749694327, + "loss": 0.1544, + "step": 37040 + }, + { + "epoch": 1.4305571643692807, + "grad_norm": 0.7789367437362671, + "learning_rate": 0.00010463209647734147, + "loss": 0.2251, + "step": 37050 + }, + { + "epoch": 1.4309432796633075, + "grad_norm": 1.1705437898635864, + "learning_rate": 0.0001046063554577397, + "loss": 0.1567, + "step": 37060 + }, + { + "epoch": 1.4313293949573342, + "grad_norm": 0.13087430596351624, + "learning_rate": 0.00010458061443813789, + "loss": 0.3237, + "step": 37070 + }, + { + "epoch": 1.431715510251361, + "grad_norm": 1.0658761262893677, + "learning_rate": 0.00010455487341853612, + "loss": 0.24, + "step": 37080 + }, + { + "epoch": 1.4321016255453878, + "grad_norm": 1.686922311782837, + "learning_rate": 0.00010452913239893433, + "loss": 0.1911, + "step": 37090 + }, + { + "epoch": 1.4324877408394148, + "grad_norm": 2.010221481323242, + "learning_rate": 0.00010450339137933253, + "loss": 0.4569, + "step": 37100 + }, + { + "epoch": 1.4328738561334413, + "grad_norm": 0.8007562160491943, + "learning_rate": 0.00010447765035973076, + "loss": 0.1198, + "step": 37110 + }, + { + "epoch": 1.4332599714274683, + "grad_norm": 0.5455211997032166, + "learning_rate": 0.00010445190934012896, + "loss": 0.2179, + "step": 37120 + }, + { + "epoch": 1.433646086721495, + "grad_norm": 1.4133542776107788, + "learning_rate": 0.00010442616832052719, + "loss": 0.45, + "step": 37130 + }, + { + "epoch": 1.4340322020155218, + "grad_norm": 1.6218222379684448, + "learning_rate": 0.00010440042730092539, + "loss": 0.19, + "step": 37140 + }, + { + "epoch": 1.4344183173095486, + "grad_norm": 0.6425970196723938, + "learning_rate": 0.00010437468628132361, + "loss": 0.2345, + "step": 37150 + }, + { + "epoch": 1.4348044326035754, + "grad_norm": 0.4344918131828308, + "learning_rate": 0.00010434894526172183, + "loss": 0.2532, + "step": 37160 + }, + { + "epoch": 1.4351905478976021, + "grad_norm": 0.6628998517990112, + "learning_rate": 0.00010432320424212003, + "loss": 0.2281, + "step": 37170 + }, + { + "epoch": 1.435576663191629, + "grad_norm": 0.8695842027664185, + "learning_rate": 0.00010429746322251825, + "loss": 0.2796, + "step": 37180 + }, + { + "epoch": 1.435962778485656, + "grad_norm": 0.16481854021549225, + "learning_rate": 0.00010427172220291645, + "loss": 0.2333, + "step": 37190 + }, + { + "epoch": 1.4363488937796827, + "grad_norm": 0.7194454073905945, + "learning_rate": 0.00010424598118331468, + "loss": 0.1413, + "step": 37200 + }, + { + "epoch": 1.4367350090737094, + "grad_norm": 4.845508575439453, + "learning_rate": 0.00010422024016371288, + "loss": 0.3944, + "step": 37210 + }, + { + "epoch": 1.4371211243677362, + "grad_norm": 2.6386618614196777, + "learning_rate": 0.0001041944991441111, + "loss": 0.3037, + "step": 37220 + }, + { + "epoch": 1.437507239661763, + "grad_norm": 0.4089922308921814, + "learning_rate": 0.00010416875812450932, + "loss": 0.2985, + "step": 37230 + }, + { + "epoch": 1.4378933549557897, + "grad_norm": 1.456944465637207, + "learning_rate": 0.00010414301710490752, + "loss": 0.3031, + "step": 37240 + }, + { + "epoch": 1.4382794702498165, + "grad_norm": 1.301829218864441, + "learning_rate": 0.00010411727608530575, + "loss": 0.2578, + "step": 37250 + }, + { + "epoch": 1.4386655855438435, + "grad_norm": 1.2072703838348389, + "learning_rate": 0.00010409153506570395, + "loss": 0.391, + "step": 37260 + }, + { + "epoch": 1.4390517008378703, + "grad_norm": 0.5538531541824341, + "learning_rate": 0.00010406579404610217, + "loss": 0.1867, + "step": 37270 + }, + { + "epoch": 1.439437816131897, + "grad_norm": 1.0898678302764893, + "learning_rate": 0.0001040400530265004, + "loss": 0.2112, + "step": 37280 + }, + { + "epoch": 1.4398239314259238, + "grad_norm": 1.5769239664077759, + "learning_rate": 0.0001040143120068986, + "loss": 0.3121, + "step": 37290 + }, + { + "epoch": 1.4402100467199506, + "grad_norm": 0.3964422345161438, + "learning_rate": 0.00010398857098729681, + "loss": 0.196, + "step": 37300 + }, + { + "epoch": 1.4405961620139773, + "grad_norm": 1.0268182754516602, + "learning_rate": 0.00010396282996769501, + "loss": 0.1785, + "step": 37310 + }, + { + "epoch": 1.440982277308004, + "grad_norm": 1.750826358795166, + "learning_rate": 0.00010393708894809324, + "loss": 0.3654, + "step": 37320 + }, + { + "epoch": 1.441368392602031, + "grad_norm": 1.1231745481491089, + "learning_rate": 0.00010391134792849144, + "loss": 0.2594, + "step": 37330 + }, + { + "epoch": 1.4417545078960576, + "grad_norm": 0.2897786498069763, + "learning_rate": 0.00010388560690888967, + "loss": 0.3483, + "step": 37340 + }, + { + "epoch": 1.4421406231900846, + "grad_norm": 0.07170752435922623, + "learning_rate": 0.00010385986588928789, + "loss": 0.1462, + "step": 37350 + }, + { + "epoch": 1.4425267384841114, + "grad_norm": 3.326099395751953, + "learning_rate": 0.00010383412486968609, + "loss": 0.3785, + "step": 37360 + }, + { + "epoch": 1.4429128537781382, + "grad_norm": 0.5063263773918152, + "learning_rate": 0.0001038083838500843, + "loss": 0.2974, + "step": 37370 + }, + { + "epoch": 1.443298969072165, + "grad_norm": 1.160088062286377, + "learning_rate": 0.00010378264283048252, + "loss": 0.3003, + "step": 37380 + }, + { + "epoch": 1.4436850843661917, + "grad_norm": 0.7043284177780151, + "learning_rate": 0.00010375690181088073, + "loss": 0.3193, + "step": 37390 + }, + { + "epoch": 1.4440711996602185, + "grad_norm": 0.2916620373725891, + "learning_rate": 0.00010373116079127896, + "loss": 0.235, + "step": 37400 + }, + { + "epoch": 1.4444573149542452, + "grad_norm": 2.1940219402313232, + "learning_rate": 0.00010370541977167716, + "loss": 0.2657, + "step": 37410 + }, + { + "epoch": 1.4448434302482722, + "grad_norm": 1.849794626235962, + "learning_rate": 0.00010367967875207539, + "loss": 0.3121, + "step": 37420 + }, + { + "epoch": 1.445229545542299, + "grad_norm": 1.1139589548110962, + "learning_rate": 0.00010365393773247359, + "loss": 0.2179, + "step": 37430 + }, + { + "epoch": 1.4456156608363258, + "grad_norm": 0.917142927646637, + "learning_rate": 0.0001036281967128718, + "loss": 0.2406, + "step": 37440 + }, + { + "epoch": 1.4460017761303525, + "grad_norm": 1.6973673105239868, + "learning_rate": 0.00010360245569327001, + "loss": 0.3236, + "step": 37450 + }, + { + "epoch": 1.4463878914243793, + "grad_norm": 0.3979933559894562, + "learning_rate": 0.00010357671467366823, + "loss": 0.1477, + "step": 37460 + }, + { + "epoch": 1.446774006718406, + "grad_norm": 0.8938451409339905, + "learning_rate": 0.00010355097365406645, + "loss": 0.116, + "step": 37470 + }, + { + "epoch": 1.4471601220124328, + "grad_norm": 1.068787693977356, + "learning_rate": 0.00010352523263446465, + "loss": 0.2342, + "step": 37480 + }, + { + "epoch": 1.4475462373064598, + "grad_norm": 1.1985591650009155, + "learning_rate": 0.00010349949161486288, + "loss": 0.2423, + "step": 37490 + }, + { + "epoch": 1.4479323526004866, + "grad_norm": 1.0383973121643066, + "learning_rate": 0.00010347375059526108, + "loss": 0.3059, + "step": 37500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.2835920896e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f75a6639e0d198dcce68d80edfb512a0335cba5a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf7343f0aba42e4f726328b5a8476b6f94c2a1097616f16bf8aaf1e8e5b6d043 +size 290979840 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..68e3970056e7d6ac99532d8170977d856db5cbe4 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c6cf18176dd07e5c2b873ce7b3929022c31d718ce300f148866bb5307b8185f +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..67c235aebc361c33435bb8f7e6dbd24d0ebdfcdc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..e244bebecae58489ebfeeebacdfed59b9c832a3e Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..073d04768c92f72ac7882d5e6c32e450e5b0752a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/trainer_state.json @@ -0,0 +1,26634 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.4672381173018263, + "eval_steps": 500, + "global_step": 38000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + }, + { + "epoch": 1.274566585582455, + "grad_norm": 0.5019121766090393, + "learning_rate": 0.0001150314683964632, + "loss": 0.1857, + "step": 33010 + }, + { + "epoch": 1.2749527008764816, + "grad_norm": 0.258952260017395, + "learning_rate": 0.0001150057273768614, + "loss": 0.11, + "step": 33020 + }, + { + "epoch": 1.2753388161705086, + "grad_norm": 0.8540348410606384, + "learning_rate": 0.00011497998635725962, + "loss": 0.1852, + "step": 33030 + }, + { + "epoch": 1.2757249314645354, + "grad_norm": 0.08689398318529129, + "learning_rate": 0.00011495424533765783, + "loss": 0.2225, + "step": 33040 + }, + { + "epoch": 1.2761110467585621, + "grad_norm": 0.42253822088241577, + "learning_rate": 0.00011492850431805605, + "loss": 0.0751, + "step": 33050 + }, + { + "epoch": 1.276497162052589, + "grad_norm": 1.2964017391204834, + "learning_rate": 0.00011490276329845426, + "loss": 0.2384, + "step": 33060 + }, + { + "epoch": 1.2768832773466157, + "grad_norm": 0.5337836146354675, + "learning_rate": 0.00011487702227885246, + "loss": 0.1415, + "step": 33070 + }, + { + "epoch": 1.2772693926406424, + "grad_norm": 2.7771682739257812, + "learning_rate": 0.00011485128125925069, + "loss": 0.378, + "step": 33080 + }, + { + "epoch": 1.2776555079346692, + "grad_norm": 1.5107232332229614, + "learning_rate": 0.00011482554023964889, + "loss": 0.2482, + "step": 33090 + }, + { + "epoch": 1.2780416232286962, + "grad_norm": 0.6885499358177185, + "learning_rate": 0.00011479979922004711, + "loss": 0.2122, + "step": 33100 + }, + { + "epoch": 1.278427738522723, + "grad_norm": 0.9016557335853577, + "learning_rate": 0.00011477405820044533, + "loss": 0.2841, + "step": 33110 + }, + { + "epoch": 1.2788138538167497, + "grad_norm": 1.9532525539398193, + "learning_rate": 0.00011474831718084354, + "loss": 0.2281, + "step": 33120 + }, + { + "epoch": 1.2791999691107765, + "grad_norm": 2.1078782081604004, + "learning_rate": 0.00011472257616124175, + "loss": 0.2836, + "step": 33130 + }, + { + "epoch": 1.2795860844048033, + "grad_norm": 0.19830390810966492, + "learning_rate": 0.00011469683514163995, + "loss": 0.388, + "step": 33140 + }, + { + "epoch": 1.27997219969883, + "grad_norm": 0.17538850009441376, + "learning_rate": 0.00011467109412203818, + "loss": 0.3274, + "step": 33150 + }, + { + "epoch": 1.2803583149928568, + "grad_norm": 0.7402139902114868, + "learning_rate": 0.00011464535310243641, + "loss": 0.1979, + "step": 33160 + }, + { + "epoch": 1.2807444302868838, + "grad_norm": 0.2097146362066269, + "learning_rate": 0.00011461961208283461, + "loss": 0.2464, + "step": 33170 + }, + { + "epoch": 1.2811305455809103, + "grad_norm": 1.2441083192825317, + "learning_rate": 0.00011459387106323283, + "loss": 0.266, + "step": 33180 + }, + { + "epoch": 1.2815166608749373, + "grad_norm": 2.518852710723877, + "learning_rate": 0.00011456813004363103, + "loss": 0.253, + "step": 33190 + }, + { + "epoch": 1.281902776168964, + "grad_norm": 0.8078998327255249, + "learning_rate": 0.00011454238902402925, + "loss": 0.2361, + "step": 33200 + }, + { + "epoch": 1.2822888914629909, + "grad_norm": 1.2297371625900269, + "learning_rate": 0.00011451664800442745, + "loss": 0.1974, + "step": 33210 + }, + { + "epoch": 1.2826750067570176, + "grad_norm": 0.4303855895996094, + "learning_rate": 0.00011449090698482567, + "loss": 0.3563, + "step": 33220 + }, + { + "epoch": 1.2830611220510444, + "grad_norm": 1.3215210437774658, + "learning_rate": 0.0001144651659652239, + "loss": 0.2818, + "step": 33230 + }, + { + "epoch": 1.2834472373450712, + "grad_norm": 1.546265959739685, + "learning_rate": 0.0001144394249456221, + "loss": 0.5778, + "step": 33240 + }, + { + "epoch": 1.283833352639098, + "grad_norm": 0.8895953297615051, + "learning_rate": 0.00011441368392602033, + "loss": 0.2485, + "step": 33250 + }, + { + "epoch": 1.284219467933125, + "grad_norm": 0.7534870505332947, + "learning_rate": 0.00011438794290641853, + "loss": 0.2825, + "step": 33260 + }, + { + "epoch": 1.2846055832271517, + "grad_norm": 0.052820973098278046, + "learning_rate": 0.00011436220188681674, + "loss": 0.2191, + "step": 33270 + }, + { + "epoch": 1.2849916985211784, + "grad_norm": 0.9264475107192993, + "learning_rate": 0.00011433646086721494, + "loss": 0.181, + "step": 33280 + }, + { + "epoch": 1.2853778138152052, + "grad_norm": 0.2128441333770752, + "learning_rate": 0.00011431071984761317, + "loss": 0.1819, + "step": 33290 + }, + { + "epoch": 1.285763929109232, + "grad_norm": 0.5400950312614441, + "learning_rate": 0.0001142849788280114, + "loss": 0.4316, + "step": 33300 + }, + { + "epoch": 1.2861500444032588, + "grad_norm": 0.8033271431922913, + "learning_rate": 0.00011425923780840959, + "loss": 0.2146, + "step": 33310 + }, + { + "epoch": 1.2865361596972855, + "grad_norm": 2.012575149536133, + "learning_rate": 0.00011423349678880782, + "loss": 0.4335, + "step": 33320 + }, + { + "epoch": 1.2869222749913125, + "grad_norm": 0.7352376580238342, + "learning_rate": 0.00011420775576920602, + "loss": 0.2124, + "step": 33330 + }, + { + "epoch": 1.2873083902853393, + "grad_norm": 0.769036591053009, + "learning_rate": 0.00011418201474960423, + "loss": 0.3602, + "step": 33340 + }, + { + "epoch": 1.287694505579366, + "grad_norm": 0.250592976808548, + "learning_rate": 0.00011415627373000246, + "loss": 0.1692, + "step": 33350 + }, + { + "epoch": 1.2880806208733928, + "grad_norm": 2.43820858001709, + "learning_rate": 0.00011413053271040066, + "loss": 0.2777, + "step": 33360 + }, + { + "epoch": 1.2884667361674196, + "grad_norm": 1.3179954290390015, + "learning_rate": 0.00011410479169079889, + "loss": 0.1794, + "step": 33370 + }, + { + "epoch": 1.2888528514614463, + "grad_norm": 1.0040466785430908, + "learning_rate": 0.00011407905067119709, + "loss": 0.3037, + "step": 33380 + }, + { + "epoch": 1.2892389667554731, + "grad_norm": 5.296288013458252, + "learning_rate": 0.00011405330965159531, + "loss": 0.2904, + "step": 33390 + }, + { + "epoch": 1.2896250820495, + "grad_norm": 0.4267273247241974, + "learning_rate": 0.00011402756863199351, + "loss": 0.2263, + "step": 33400 + }, + { + "epoch": 1.2900111973435266, + "grad_norm": 0.8817713260650635, + "learning_rate": 0.00011400182761239173, + "loss": 0.2715, + "step": 33410 + }, + { + "epoch": 1.2903973126375536, + "grad_norm": 2.7891275882720947, + "learning_rate": 0.00011397608659278995, + "loss": 0.2781, + "step": 33420 + }, + { + "epoch": 1.2907834279315804, + "grad_norm": 0.3013952374458313, + "learning_rate": 0.00011395034557318815, + "loss": 0.2563, + "step": 33430 + }, + { + "epoch": 1.2911695432256072, + "grad_norm": 1.766413927078247, + "learning_rate": 0.00011392460455358638, + "loss": 0.1854, + "step": 33440 + }, + { + "epoch": 1.291555658519634, + "grad_norm": 0.25331103801727295, + "learning_rate": 0.00011389886353398458, + "loss": 0.1206, + "step": 33450 + }, + { + "epoch": 1.2919417738136607, + "grad_norm": 0.036400288343429565, + "learning_rate": 0.0001138731225143828, + "loss": 0.4707, + "step": 33460 + }, + { + "epoch": 1.2923278891076875, + "grad_norm": 1.5620888471603394, + "learning_rate": 0.00011384738149478102, + "loss": 0.3312, + "step": 33470 + }, + { + "epoch": 1.2927140044017142, + "grad_norm": 0.6670392155647278, + "learning_rate": 0.00011382164047517922, + "loss": 0.2341, + "step": 33480 + }, + { + "epoch": 1.2931001196957412, + "grad_norm": 2.3108737468719482, + "learning_rate": 0.00011379589945557745, + "loss": 0.3843, + "step": 33490 + }, + { + "epoch": 1.293486234989768, + "grad_norm": 0.8025147318840027, + "learning_rate": 0.00011377015843597565, + "loss": 0.1982, + "step": 33500 + }, + { + "epoch": 1.2938723502837948, + "grad_norm": 1.7835719585418701, + "learning_rate": 0.00011374441741637387, + "loss": 0.3285, + "step": 33510 + }, + { + "epoch": 1.2942584655778215, + "grad_norm": 2.041508913040161, + "learning_rate": 0.00011371867639677207, + "loss": 0.2044, + "step": 33520 + }, + { + "epoch": 1.2946445808718483, + "grad_norm": 1.103378415107727, + "learning_rate": 0.0001136929353771703, + "loss": 0.1682, + "step": 33530 + }, + { + "epoch": 1.295030696165875, + "grad_norm": 0.057376351207494736, + "learning_rate": 0.00011366719435756851, + "loss": 0.1642, + "step": 33540 + }, + { + "epoch": 1.2954168114599018, + "grad_norm": 0.6539410948753357, + "learning_rate": 0.00011364145333796671, + "loss": 0.1549, + "step": 33550 + }, + { + "epoch": 1.2958029267539288, + "grad_norm": 1.250543236732483, + "learning_rate": 0.00011361571231836494, + "loss": 0.3764, + "step": 33560 + }, + { + "epoch": 1.2961890420479556, + "grad_norm": 0.23697887361049652, + "learning_rate": 0.00011358997129876314, + "loss": 0.3999, + "step": 33570 + }, + { + "epoch": 1.2965751573419824, + "grad_norm": 0.9318505525588989, + "learning_rate": 0.00011356423027916137, + "loss": 0.4156, + "step": 33580 + }, + { + "epoch": 1.2969612726360091, + "grad_norm": 1.3910777568817139, + "learning_rate": 0.00011353848925955957, + "loss": 0.3455, + "step": 33590 + }, + { + "epoch": 1.297347387930036, + "grad_norm": 1.6764451265335083, + "learning_rate": 0.00011351274823995779, + "loss": 0.1884, + "step": 33600 + }, + { + "epoch": 1.2977335032240627, + "grad_norm": 0.9300051927566528, + "learning_rate": 0.000113487007220356, + "loss": 0.119, + "step": 33610 + }, + { + "epoch": 1.2981196185180894, + "grad_norm": 2.447462558746338, + "learning_rate": 0.00011346126620075422, + "loss": 0.4403, + "step": 33620 + }, + { + "epoch": 1.2985057338121164, + "grad_norm": 1.216407060623169, + "learning_rate": 0.00011343552518115243, + "loss": 0.2415, + "step": 33630 + }, + { + "epoch": 1.298891849106143, + "grad_norm": 2.968648910522461, + "learning_rate": 0.00011340978416155063, + "loss": 0.2899, + "step": 33640 + }, + { + "epoch": 1.29927796440017, + "grad_norm": 0.6649970412254333, + "learning_rate": 0.00011338404314194886, + "loss": 0.3809, + "step": 33650 + }, + { + "epoch": 1.2996640796941967, + "grad_norm": 1.7277917861938477, + "learning_rate": 0.00011335830212234709, + "loss": 0.3308, + "step": 33660 + }, + { + "epoch": 1.3000501949882235, + "grad_norm": 1.3269709348678589, + "learning_rate": 0.00011333256110274529, + "loss": 0.3682, + "step": 33670 + }, + { + "epoch": 1.3004363102822503, + "grad_norm": 0.20609407126903534, + "learning_rate": 0.0001133068200831435, + "loss": 0.1379, + "step": 33680 + }, + { + "epoch": 1.300822425576277, + "grad_norm": 0.6592215299606323, + "learning_rate": 0.00011328107906354171, + "loss": 0.2746, + "step": 33690 + }, + { + "epoch": 1.3012085408703038, + "grad_norm": 1.903635859489441, + "learning_rate": 0.00011325533804393993, + "loss": 0.4729, + "step": 33700 + }, + { + "epoch": 1.3015946561643306, + "grad_norm": 0.8432504534721375, + "learning_rate": 0.00011322959702433813, + "loss": 0.2835, + "step": 33710 + }, + { + "epoch": 1.3019807714583576, + "grad_norm": 0.9862542152404785, + "learning_rate": 0.00011320385600473635, + "loss": 0.1823, + "step": 33720 + }, + { + "epoch": 1.3023668867523843, + "grad_norm": 3.845738649368286, + "learning_rate": 0.00011317811498513458, + "loss": 0.2542, + "step": 33730 + }, + { + "epoch": 1.302753002046411, + "grad_norm": 0.6317747235298157, + "learning_rate": 0.00011315237396553278, + "loss": 0.22, + "step": 33740 + }, + { + "epoch": 1.3031391173404379, + "grad_norm": 2.5221354961395264, + "learning_rate": 0.000113126632945931, + "loss": 0.2253, + "step": 33750 + }, + { + "epoch": 1.3035252326344646, + "grad_norm": 1.3326247930526733, + "learning_rate": 0.0001131008919263292, + "loss": 0.2021, + "step": 33760 + }, + { + "epoch": 1.3039113479284914, + "grad_norm": 1.148047685623169, + "learning_rate": 0.00011307515090672742, + "loss": 0.3987, + "step": 33770 + }, + { + "epoch": 1.3042974632225182, + "grad_norm": 0.19721268117427826, + "learning_rate": 0.00011304940988712562, + "loss": 0.2642, + "step": 33780 + }, + { + "epoch": 1.3046835785165452, + "grad_norm": 1.4060617685317993, + "learning_rate": 0.00011302366886752385, + "loss": 0.2736, + "step": 33790 + }, + { + "epoch": 1.305069693810572, + "grad_norm": 1.0736548900604248, + "learning_rate": 0.00011299792784792207, + "loss": 0.2448, + "step": 33800 + }, + { + "epoch": 1.3054558091045987, + "grad_norm": 4.352476119995117, + "learning_rate": 0.00011297218682832027, + "loss": 0.383, + "step": 33810 + }, + { + "epoch": 1.3058419243986255, + "grad_norm": 0.2249228060245514, + "learning_rate": 0.0001129464458087185, + "loss": 0.14, + "step": 33820 + }, + { + "epoch": 1.3062280396926522, + "grad_norm": 0.4820781946182251, + "learning_rate": 0.0001129207047891167, + "loss": 0.248, + "step": 33830 + }, + { + "epoch": 1.306614154986679, + "grad_norm": 2.2983391284942627, + "learning_rate": 0.00011289496376951491, + "loss": 0.2608, + "step": 33840 + }, + { + "epoch": 1.3070002702807058, + "grad_norm": 1.3315671682357788, + "learning_rate": 0.00011286922274991314, + "loss": 0.1727, + "step": 33850 + }, + { + "epoch": 1.3073863855747327, + "grad_norm": 2.060299873352051, + "learning_rate": 0.00011284348173031134, + "loss": 0.3097, + "step": 33860 + }, + { + "epoch": 1.3077725008687593, + "grad_norm": 2.096285581588745, + "learning_rate": 0.00011281774071070957, + "loss": 0.2305, + "step": 33870 + }, + { + "epoch": 1.3081586161627863, + "grad_norm": 0.4997636675834656, + "learning_rate": 0.00011279199969110777, + "loss": 0.1993, + "step": 33880 + }, + { + "epoch": 1.308544731456813, + "grad_norm": 0.58636474609375, + "learning_rate": 0.00011276625867150599, + "loss": 0.1873, + "step": 33890 + }, + { + "epoch": 1.3089308467508398, + "grad_norm": 0.9128592610359192, + "learning_rate": 0.00011274051765190419, + "loss": 0.1885, + "step": 33900 + }, + { + "epoch": 1.3093169620448666, + "grad_norm": 2.228043794631958, + "learning_rate": 0.0001127147766323024, + "loss": 0.3649, + "step": 33910 + }, + { + "epoch": 1.3097030773388934, + "grad_norm": 1.069002389907837, + "learning_rate": 0.00011268903561270063, + "loss": 0.5454, + "step": 33920 + }, + { + "epoch": 1.3100891926329201, + "grad_norm": 0.6207597851753235, + "learning_rate": 0.00011266329459309883, + "loss": 0.2329, + "step": 33930 + }, + { + "epoch": 1.3104753079269469, + "grad_norm": 1.262247920036316, + "learning_rate": 0.00011263755357349706, + "loss": 0.3437, + "step": 33940 + }, + { + "epoch": 1.3108614232209739, + "grad_norm": 1.7429994344711304, + "learning_rate": 0.00011261181255389526, + "loss": 0.228, + "step": 33950 + }, + { + "epoch": 1.3112475385150006, + "grad_norm": 0.646900475025177, + "learning_rate": 0.00011258607153429349, + "loss": 0.3739, + "step": 33960 + }, + { + "epoch": 1.3116336538090274, + "grad_norm": 1.8228782415390015, + "learning_rate": 0.0001125603305146917, + "loss": 0.2325, + "step": 33970 + }, + { + "epoch": 1.3120197691030542, + "grad_norm": 3.539228916168213, + "learning_rate": 0.0001125345894950899, + "loss": 0.194, + "step": 33980 + }, + { + "epoch": 1.312405884397081, + "grad_norm": 1.2801135778427124, + "learning_rate": 0.00011250884847548813, + "loss": 0.3069, + "step": 33990 + }, + { + "epoch": 1.3127919996911077, + "grad_norm": 3.6265695095062256, + "learning_rate": 0.00011248310745588633, + "loss": 0.3113, + "step": 34000 + }, + { + "epoch": 1.3131781149851345, + "grad_norm": 0.07370063662528992, + "learning_rate": 0.00011245736643628455, + "loss": 0.1449, + "step": 34010 + }, + { + "epoch": 1.3135642302791615, + "grad_norm": 1.0295637845993042, + "learning_rate": 0.00011243162541668275, + "loss": 0.22, + "step": 34020 + }, + { + "epoch": 1.313950345573188, + "grad_norm": 0.8803662061691284, + "learning_rate": 0.00011240588439708098, + "loss": 0.1368, + "step": 34030 + }, + { + "epoch": 1.314336460867215, + "grad_norm": 1.6597707271575928, + "learning_rate": 0.00011238014337747919, + "loss": 0.3038, + "step": 34040 + }, + { + "epoch": 1.3147225761612418, + "grad_norm": 2.115492343902588, + "learning_rate": 0.00011235440235787739, + "loss": 0.1754, + "step": 34050 + }, + { + "epoch": 1.3151086914552685, + "grad_norm": 0.8143919706344604, + "learning_rate": 0.00011232866133827562, + "loss": 0.3764, + "step": 34060 + }, + { + "epoch": 1.3154948067492953, + "grad_norm": 0.14369767904281616, + "learning_rate": 0.00011230292031867382, + "loss": 0.1742, + "step": 34070 + }, + { + "epoch": 1.315880922043322, + "grad_norm": 1.0129845142364502, + "learning_rate": 0.00011227717929907205, + "loss": 0.1458, + "step": 34080 + }, + { + "epoch": 1.316267037337349, + "grad_norm": 2.7300291061401367, + "learning_rate": 0.00011225143827947025, + "loss": 0.3939, + "step": 34090 + }, + { + "epoch": 1.3166531526313756, + "grad_norm": 0.20205609500408173, + "learning_rate": 0.00011222569725986847, + "loss": 0.29, + "step": 34100 + }, + { + "epoch": 1.3170392679254026, + "grad_norm": 1.8928464651107788, + "learning_rate": 0.00011219995624026669, + "loss": 0.1742, + "step": 34110 + }, + { + "epoch": 1.3174253832194294, + "grad_norm": 0.2639687955379486, + "learning_rate": 0.00011217421522066488, + "loss": 0.1745, + "step": 34120 + }, + { + "epoch": 1.3178114985134561, + "grad_norm": 0.5906389355659485, + "learning_rate": 0.00011214847420106311, + "loss": 0.2134, + "step": 34130 + }, + { + "epoch": 1.318197613807483, + "grad_norm": 0.9190629720687866, + "learning_rate": 0.00011212273318146131, + "loss": 0.2547, + "step": 34140 + }, + { + "epoch": 1.3185837291015097, + "grad_norm": 0.5749151110649109, + "learning_rate": 0.00011209699216185954, + "loss": 0.1688, + "step": 34150 + }, + { + "epoch": 1.3189698443955364, + "grad_norm": 0.82295823097229, + "learning_rate": 0.00011207125114225777, + "loss": 0.2884, + "step": 34160 + }, + { + "epoch": 1.3193559596895632, + "grad_norm": 0.07816460728645325, + "learning_rate": 0.00011204551012265597, + "loss": 0.2418, + "step": 34170 + }, + { + "epoch": 1.3197420749835902, + "grad_norm": 0.6417407393455505, + "learning_rate": 0.00011201976910305418, + "loss": 0.2557, + "step": 34180 + }, + { + "epoch": 1.320128190277617, + "grad_norm": 6.093267440795898, + "learning_rate": 0.00011199402808345238, + "loss": 0.3088, + "step": 34190 + }, + { + "epoch": 1.3205143055716437, + "grad_norm": 1.8861887454986572, + "learning_rate": 0.0001119682870638506, + "loss": 0.2204, + "step": 34200 + }, + { + "epoch": 1.3209004208656705, + "grad_norm": 2.3272714614868164, + "learning_rate": 0.0001119425460442488, + "loss": 0.2236, + "step": 34210 + }, + { + "epoch": 1.3212865361596973, + "grad_norm": 0.9608810544013977, + "learning_rate": 0.00011191680502464703, + "loss": 0.1897, + "step": 34220 + }, + { + "epoch": 1.321672651453724, + "grad_norm": 1.2157350778579712, + "learning_rate": 0.00011189106400504526, + "loss": 0.1526, + "step": 34230 + }, + { + "epoch": 1.3220587667477508, + "grad_norm": 1.6684671640396118, + "learning_rate": 0.00011186532298544346, + "loss": 0.3394, + "step": 34240 + }, + { + "epoch": 1.3224448820417778, + "grad_norm": 2.0432374477386475, + "learning_rate": 0.00011183958196584167, + "loss": 0.2183, + "step": 34250 + }, + { + "epoch": 1.3228309973358043, + "grad_norm": 0.9436892867088318, + "learning_rate": 0.00011181384094623988, + "loss": 0.2947, + "step": 34260 + }, + { + "epoch": 1.3232171126298313, + "grad_norm": 0.23260092735290527, + "learning_rate": 0.0001117880999266381, + "loss": 0.114, + "step": 34270 + }, + { + "epoch": 1.323603227923858, + "grad_norm": 1.2291594743728638, + "learning_rate": 0.0001117623589070363, + "loss": 0.3145, + "step": 34280 + }, + { + "epoch": 1.3239893432178849, + "grad_norm": 0.41411107778549194, + "learning_rate": 0.00011173661788743452, + "loss": 0.2937, + "step": 34290 + }, + { + "epoch": 1.3243754585119116, + "grad_norm": 2.354405164718628, + "learning_rate": 0.00011171087686783275, + "loss": 0.3933, + "step": 34300 + }, + { + "epoch": 1.3247615738059384, + "grad_norm": 2.6997978687286377, + "learning_rate": 0.00011168513584823095, + "loss": 0.1494, + "step": 34310 + }, + { + "epoch": 1.3251476890999654, + "grad_norm": 2.8430919647216797, + "learning_rate": 0.00011165939482862916, + "loss": 0.2869, + "step": 34320 + }, + { + "epoch": 1.325533804393992, + "grad_norm": 1.1737356185913086, + "learning_rate": 0.00011163365380902738, + "loss": 0.2792, + "step": 34330 + }, + { + "epoch": 1.325919919688019, + "grad_norm": 4.123973846435547, + "learning_rate": 0.00011160791278942559, + "loss": 0.5211, + "step": 34340 + }, + { + "epoch": 1.3263060349820457, + "grad_norm": 0.8862038850784302, + "learning_rate": 0.00011158217176982382, + "loss": 0.2976, + "step": 34350 + }, + { + "epoch": 1.3266921502760725, + "grad_norm": 1.8690590858459473, + "learning_rate": 0.00011155643075022202, + "loss": 0.2485, + "step": 34360 + }, + { + "epoch": 1.3270782655700992, + "grad_norm": 2.885589599609375, + "learning_rate": 0.00011153068973062024, + "loss": 0.3084, + "step": 34370 + }, + { + "epoch": 1.327464380864126, + "grad_norm": 0.9898788928985596, + "learning_rate": 0.00011150494871101844, + "loss": 0.261, + "step": 34380 + }, + { + "epoch": 1.3278504961581528, + "grad_norm": 0.6879653930664062, + "learning_rate": 0.00011147920769141667, + "loss": 0.2082, + "step": 34390 + }, + { + "epoch": 1.3282366114521795, + "grad_norm": 1.2619003057479858, + "learning_rate": 0.00011145346667181487, + "loss": 0.2402, + "step": 34400 + }, + { + "epoch": 1.3286227267462065, + "grad_norm": 1.1212007999420166, + "learning_rate": 0.00011142772565221308, + "loss": 0.3525, + "step": 34410 + }, + { + "epoch": 1.3290088420402333, + "grad_norm": 1.8431956768035889, + "learning_rate": 0.00011140198463261131, + "loss": 0.2212, + "step": 34420 + }, + { + "epoch": 1.32939495733426, + "grad_norm": 0.6185423731803894, + "learning_rate": 0.00011137624361300951, + "loss": 0.2455, + "step": 34430 + }, + { + "epoch": 1.3297810726282868, + "grad_norm": 2.3791301250457764, + "learning_rate": 0.00011135050259340774, + "loss": 0.1763, + "step": 34440 + }, + { + "epoch": 1.3301671879223136, + "grad_norm": 0.4928603768348694, + "learning_rate": 0.00011132476157380594, + "loss": 0.2381, + "step": 34450 + }, + { + "epoch": 1.3305533032163404, + "grad_norm": 1.5636029243469238, + "learning_rate": 0.00011129902055420416, + "loss": 0.1368, + "step": 34460 + }, + { + "epoch": 1.3309394185103671, + "grad_norm": 0.9425283074378967, + "learning_rate": 0.00011127327953460238, + "loss": 0.2825, + "step": 34470 + }, + { + "epoch": 1.3313255338043941, + "grad_norm": 1.2257115840911865, + "learning_rate": 0.00011124753851500058, + "loss": 0.2547, + "step": 34480 + }, + { + "epoch": 1.3317116490984207, + "grad_norm": 0.9416170716285706, + "learning_rate": 0.0001112217974953988, + "loss": 0.2766, + "step": 34490 + }, + { + "epoch": 1.3320977643924476, + "grad_norm": 0.5123847126960754, + "learning_rate": 0.000111196056475797, + "loss": 0.4733, + "step": 34500 + }, + { + "epoch": 1.3324838796864744, + "grad_norm": 1.5581384897232056, + "learning_rate": 0.00011117031545619523, + "loss": 0.1597, + "step": 34510 + }, + { + "epoch": 1.3328699949805012, + "grad_norm": 2.377333879470825, + "learning_rate": 0.00011114457443659343, + "loss": 0.209, + "step": 34520 + }, + { + "epoch": 1.333256110274528, + "grad_norm": 1.7840913534164429, + "learning_rate": 0.00011111883341699166, + "loss": 0.1759, + "step": 34530 + }, + { + "epoch": 1.3336422255685547, + "grad_norm": 1.1825993061065674, + "learning_rate": 0.00011109309239738987, + "loss": 0.2464, + "step": 34540 + }, + { + "epoch": 1.3340283408625815, + "grad_norm": 1.8859659433364868, + "learning_rate": 0.00011106735137778807, + "loss": 0.3539, + "step": 34550 + }, + { + "epoch": 1.3344144561566083, + "grad_norm": 1.9698175191879272, + "learning_rate": 0.0001110416103581863, + "loss": 0.3301, + "step": 34560 + }, + { + "epoch": 1.3348005714506352, + "grad_norm": 0.7649385333061218, + "learning_rate": 0.0001110158693385845, + "loss": 0.232, + "step": 34570 + }, + { + "epoch": 1.335186686744662, + "grad_norm": 0.56386399269104, + "learning_rate": 0.00011099012831898272, + "loss": 0.3425, + "step": 34580 + }, + { + "epoch": 1.3355728020386888, + "grad_norm": 2.956003189086914, + "learning_rate": 0.00011096438729938092, + "loss": 0.1518, + "step": 34590 + }, + { + "epoch": 1.3359589173327155, + "grad_norm": 2.612029552459717, + "learning_rate": 0.00011093864627977915, + "loss": 0.2765, + "step": 34600 + }, + { + "epoch": 1.3363450326267423, + "grad_norm": 0.9674397706985474, + "learning_rate": 0.00011091290526017736, + "loss": 0.303, + "step": 34610 + }, + { + "epoch": 1.336731147920769, + "grad_norm": 0.9578921794891357, + "learning_rate": 0.00011088716424057556, + "loss": 0.1405, + "step": 34620 + }, + { + "epoch": 1.3371172632147958, + "grad_norm": 2.168065071105957, + "learning_rate": 0.00011086142322097379, + "loss": 0.1914, + "step": 34630 + }, + { + "epoch": 1.3375033785088228, + "grad_norm": 1.3166526556015015, + "learning_rate": 0.00011083568220137199, + "loss": 0.4134, + "step": 34640 + }, + { + "epoch": 1.3378894938028496, + "grad_norm": 0.9082283973693848, + "learning_rate": 0.00011080994118177022, + "loss": 0.2693, + "step": 34650 + }, + { + "epoch": 1.3382756090968764, + "grad_norm": 2.203007698059082, + "learning_rate": 0.00011078420016216844, + "loss": 0.1847, + "step": 34660 + }, + { + "epoch": 1.3386617243909031, + "grad_norm": 0.8101674914360046, + "learning_rate": 0.00011075845914256664, + "loss": 0.3111, + "step": 34670 + }, + { + "epoch": 1.33904783968493, + "grad_norm": 1.9545695781707764, + "learning_rate": 0.00011073271812296486, + "loss": 0.3761, + "step": 34680 + }, + { + "epoch": 1.3394339549789567, + "grad_norm": 1.547581672668457, + "learning_rate": 0.00011070697710336306, + "loss": 0.2374, + "step": 34690 + }, + { + "epoch": 1.3398200702729834, + "grad_norm": 3.3519034385681152, + "learning_rate": 0.00011068123608376128, + "loss": 0.1957, + "step": 34700 + }, + { + "epoch": 1.3402061855670104, + "grad_norm": 1.5508599281311035, + "learning_rate": 0.00011065549506415948, + "loss": 0.4171, + "step": 34710 + }, + { + "epoch": 1.340592300861037, + "grad_norm": 1.8547546863555908, + "learning_rate": 0.00011062975404455771, + "loss": 0.1872, + "step": 34720 + }, + { + "epoch": 1.340978416155064, + "grad_norm": 1.4600756168365479, + "learning_rate": 0.00011060401302495594, + "loss": 0.3515, + "step": 34730 + }, + { + "epoch": 1.3413645314490907, + "grad_norm": 0.05774044618010521, + "learning_rate": 0.00011057827200535414, + "loss": 0.1604, + "step": 34740 + }, + { + "epoch": 1.3417506467431175, + "grad_norm": 2.8793342113494873, + "learning_rate": 0.00011055253098575235, + "loss": 0.3095, + "step": 34750 + }, + { + "epoch": 1.3421367620371443, + "grad_norm": 2.241042375564575, + "learning_rate": 0.00011052678996615055, + "loss": 0.2511, + "step": 34760 + }, + { + "epoch": 1.342522877331171, + "grad_norm": 1.9320632219314575, + "learning_rate": 0.00011050104894654878, + "loss": 0.4493, + "step": 34770 + }, + { + "epoch": 1.3429089926251978, + "grad_norm": 1.6483882665634155, + "learning_rate": 0.000110475307926947, + "loss": 0.217, + "step": 34780 + }, + { + "epoch": 1.3432951079192246, + "grad_norm": 0.9635765552520752, + "learning_rate": 0.0001104495669073452, + "loss": 0.5458, + "step": 34790 + }, + { + "epoch": 1.3436812232132516, + "grad_norm": 1.2436567544937134, + "learning_rate": 0.00011042382588774343, + "loss": 0.2857, + "step": 34800 + }, + { + "epoch": 1.3440673385072783, + "grad_norm": 2.8082425594329834, + "learning_rate": 0.00011039808486814163, + "loss": 0.3439, + "step": 34810 + }, + { + "epoch": 1.344453453801305, + "grad_norm": 1.0430901050567627, + "learning_rate": 0.00011037234384853984, + "loss": 0.1404, + "step": 34820 + }, + { + "epoch": 1.3448395690953319, + "grad_norm": 1.7387149333953857, + "learning_rate": 0.00011034660282893806, + "loss": 0.395, + "step": 34830 + }, + { + "epoch": 1.3452256843893586, + "grad_norm": 1.2713748216629028, + "learning_rate": 0.00011032086180933627, + "loss": 0.29, + "step": 34840 + }, + { + "epoch": 1.3456117996833854, + "grad_norm": 0.26068204641342163, + "learning_rate": 0.0001102951207897345, + "loss": 0.1814, + "step": 34850 + }, + { + "epoch": 1.3459979149774122, + "grad_norm": 2.4163243770599365, + "learning_rate": 0.0001102693797701327, + "loss": 0.2445, + "step": 34860 + }, + { + "epoch": 1.3463840302714392, + "grad_norm": 2.2439687252044678, + "learning_rate": 0.00011024363875053092, + "loss": 0.3338, + "step": 34870 + }, + { + "epoch": 1.346770145565466, + "grad_norm": 0.2822403609752655, + "learning_rate": 0.00011021789773092912, + "loss": 0.1648, + "step": 34880 + }, + { + "epoch": 1.3471562608594927, + "grad_norm": 0.07319017499685287, + "learning_rate": 0.00011019215671132734, + "loss": 0.107, + "step": 34890 + }, + { + "epoch": 1.3475423761535195, + "grad_norm": 0.9809044003486633, + "learning_rate": 0.00011016641569172555, + "loss": 0.256, + "step": 34900 + }, + { + "epoch": 1.3479284914475462, + "grad_norm": 0.5016226768493652, + "learning_rate": 0.00011014067467212376, + "loss": 0.3229, + "step": 34910 + }, + { + "epoch": 1.348314606741573, + "grad_norm": 1.3026005029678345, + "learning_rate": 0.00011011493365252199, + "loss": 0.2145, + "step": 34920 + }, + { + "epoch": 1.3487007220355998, + "grad_norm": 1.0752215385437012, + "learning_rate": 0.00011008919263292019, + "loss": 0.2355, + "step": 34930 + }, + { + "epoch": 1.3490868373296268, + "grad_norm": 2.2703003883361816, + "learning_rate": 0.00011006345161331842, + "loss": 0.2079, + "step": 34940 + }, + { + "epoch": 1.3494729526236533, + "grad_norm": 1.1323810815811157, + "learning_rate": 0.00011003771059371662, + "loss": 0.1015, + "step": 34950 + }, + { + "epoch": 1.3498590679176803, + "grad_norm": 0.10813555121421814, + "learning_rate": 0.00011001196957411484, + "loss": 0.4214, + "step": 34960 + }, + { + "epoch": 1.350245183211707, + "grad_norm": 0.07815568149089813, + "learning_rate": 0.00010998622855451306, + "loss": 0.1211, + "step": 34970 + }, + { + "epoch": 1.3506312985057338, + "grad_norm": 0.6748234629631042, + "learning_rate": 0.00010996048753491126, + "loss": 0.3508, + "step": 34980 + }, + { + "epoch": 1.3510174137997606, + "grad_norm": 1.8556997776031494, + "learning_rate": 0.00010993474651530948, + "loss": 0.2268, + "step": 34990 + }, + { + "epoch": 1.3514035290937874, + "grad_norm": 0.8696061372756958, + "learning_rate": 0.00010990900549570768, + "loss": 0.4321, + "step": 35000 + }, + { + "epoch": 1.3517896443878141, + "grad_norm": 0.42442765831947327, + "learning_rate": 0.00010988326447610591, + "loss": 0.1944, + "step": 35010 + }, + { + "epoch": 1.352175759681841, + "grad_norm": 1.0474554300308228, + "learning_rate": 0.00010985752345650411, + "loss": 0.1342, + "step": 35020 + }, + { + "epoch": 1.3525618749758679, + "grad_norm": 0.607037365436554, + "learning_rate": 0.00010983178243690234, + "loss": 0.2965, + "step": 35030 + }, + { + "epoch": 1.3529479902698947, + "grad_norm": 1.8160990476608276, + "learning_rate": 0.00010980604141730055, + "loss": 0.3192, + "step": 35040 + }, + { + "epoch": 1.3533341055639214, + "grad_norm": 2.0026509761810303, + "learning_rate": 0.00010978030039769875, + "loss": 0.3054, + "step": 35050 + }, + { + "epoch": 1.3537202208579482, + "grad_norm": 0.9203600883483887, + "learning_rate": 0.00010975455937809698, + "loss": 0.253, + "step": 35060 + }, + { + "epoch": 1.354106336151975, + "grad_norm": 0.33198195695877075, + "learning_rate": 0.00010972881835849518, + "loss": 0.3885, + "step": 35070 + }, + { + "epoch": 1.3544924514460017, + "grad_norm": 0.3201223611831665, + "learning_rate": 0.0001097030773388934, + "loss": 0.3029, + "step": 35080 + }, + { + "epoch": 1.3548785667400285, + "grad_norm": 1.2589943408966064, + "learning_rate": 0.0001096773363192916, + "loss": 0.4243, + "step": 35090 + }, + { + "epoch": 1.3552646820340555, + "grad_norm": 1.5106219053268433, + "learning_rate": 0.00010965159529968983, + "loss": 0.2585, + "step": 35100 + }, + { + "epoch": 1.3556507973280822, + "grad_norm": 1.429799199104309, + "learning_rate": 0.00010962585428008804, + "loss": 0.1961, + "step": 35110 + }, + { + "epoch": 1.356036912622109, + "grad_norm": 2.1211297512054443, + "learning_rate": 0.00010960011326048624, + "loss": 0.4057, + "step": 35120 + }, + { + "epoch": 1.3564230279161358, + "grad_norm": 2.5154731273651123, + "learning_rate": 0.00010957437224088447, + "loss": 0.3787, + "step": 35130 + }, + { + "epoch": 1.3568091432101625, + "grad_norm": 0.4914834201335907, + "learning_rate": 0.00010954863122128267, + "loss": 0.234, + "step": 35140 + }, + { + "epoch": 1.3571952585041893, + "grad_norm": 0.26685893535614014, + "learning_rate": 0.0001095228902016809, + "loss": 0.2841, + "step": 35150 + }, + { + "epoch": 1.357581373798216, + "grad_norm": 0.15462155640125275, + "learning_rate": 0.00010949714918207912, + "loss": 0.2269, + "step": 35160 + }, + { + "epoch": 1.357967489092243, + "grad_norm": 1.3887063264846802, + "learning_rate": 0.00010947140816247732, + "loss": 0.3455, + "step": 35170 + }, + { + "epoch": 1.3583536043862696, + "grad_norm": 0.786374032497406, + "learning_rate": 0.00010944566714287554, + "loss": 0.2897, + "step": 35180 + }, + { + "epoch": 1.3587397196802966, + "grad_norm": 1.100475549697876, + "learning_rate": 0.00010941992612327374, + "loss": 0.2892, + "step": 35190 + }, + { + "epoch": 1.3591258349743234, + "grad_norm": 0.7676102519035339, + "learning_rate": 0.00010939418510367196, + "loss": 0.1942, + "step": 35200 + }, + { + "epoch": 1.3595119502683501, + "grad_norm": 0.33462053537368774, + "learning_rate": 0.00010936844408407016, + "loss": 0.2872, + "step": 35210 + }, + { + "epoch": 1.359898065562377, + "grad_norm": 0.9294387698173523, + "learning_rate": 0.00010934270306446839, + "loss": 0.2617, + "step": 35220 + }, + { + "epoch": 1.3602841808564037, + "grad_norm": 0.3169979453086853, + "learning_rate": 0.00010931696204486662, + "loss": 0.2942, + "step": 35230 + }, + { + "epoch": 1.3606702961504304, + "grad_norm": 2.1339616775512695, + "learning_rate": 0.00010929122102526482, + "loss": 0.4448, + "step": 35240 + }, + { + "epoch": 1.3610564114444572, + "grad_norm": 0.9430062770843506, + "learning_rate": 0.00010926548000566303, + "loss": 0.2051, + "step": 35250 + }, + { + "epoch": 1.3614425267384842, + "grad_norm": 3.1187360286712646, + "learning_rate": 0.00010923973898606123, + "loss": 0.2274, + "step": 35260 + }, + { + "epoch": 1.361828642032511, + "grad_norm": 1.4727579355239868, + "learning_rate": 0.00010921399796645946, + "loss": 0.3757, + "step": 35270 + }, + { + "epoch": 1.3622147573265377, + "grad_norm": 2.157560348510742, + "learning_rate": 0.00010918825694685768, + "loss": 0.3096, + "step": 35280 + }, + { + "epoch": 1.3626008726205645, + "grad_norm": 0.33457377552986145, + "learning_rate": 0.00010916251592725588, + "loss": 0.1489, + "step": 35290 + }, + { + "epoch": 1.3629869879145913, + "grad_norm": 0.9005904197692871, + "learning_rate": 0.00010913677490765411, + "loss": 0.1826, + "step": 35300 + }, + { + "epoch": 1.363373103208618, + "grad_norm": 2.1222829818725586, + "learning_rate": 0.00010911103388805231, + "loss": 0.1965, + "step": 35310 + }, + { + "epoch": 1.3637592185026448, + "grad_norm": 1.3881357908248901, + "learning_rate": 0.00010908529286845052, + "loss": 0.1791, + "step": 35320 + }, + { + "epoch": 1.3641453337966718, + "grad_norm": 1.7574503421783447, + "learning_rate": 0.00010905955184884872, + "loss": 0.3316, + "step": 35330 + }, + { + "epoch": 1.3645314490906983, + "grad_norm": 0.1967727392911911, + "learning_rate": 0.00010903381082924695, + "loss": 0.2331, + "step": 35340 + }, + { + "epoch": 1.3649175643847253, + "grad_norm": 0.8974360823631287, + "learning_rate": 0.00010900806980964518, + "loss": 0.2589, + "step": 35350 + }, + { + "epoch": 1.365303679678752, + "grad_norm": 2.0996744632720947, + "learning_rate": 0.00010898232879004338, + "loss": 0.3663, + "step": 35360 + }, + { + "epoch": 1.3656897949727789, + "grad_norm": 0.5678316354751587, + "learning_rate": 0.0001089565877704416, + "loss": 0.1729, + "step": 35370 + }, + { + "epoch": 1.3660759102668056, + "grad_norm": 2.3381874561309814, + "learning_rate": 0.0001089308467508398, + "loss": 0.1615, + "step": 35380 + }, + { + "epoch": 1.3664620255608324, + "grad_norm": 1.0276836156845093, + "learning_rate": 0.00010890510573123802, + "loss": 0.3359, + "step": 35390 + }, + { + "epoch": 1.3668481408548594, + "grad_norm": 2.4374940395355225, + "learning_rate": 0.00010887936471163622, + "loss": 0.2435, + "step": 35400 + }, + { + "epoch": 1.367234256148886, + "grad_norm": 0.45221665501594543, + "learning_rate": 0.00010885362369203444, + "loss": 0.2555, + "step": 35410 + }, + { + "epoch": 1.367620371442913, + "grad_norm": 2.608090400695801, + "learning_rate": 0.00010882788267243267, + "loss": 0.2465, + "step": 35420 + }, + { + "epoch": 1.3680064867369397, + "grad_norm": 1.4186642169952393, + "learning_rate": 0.00010880214165283087, + "loss": 0.1674, + "step": 35430 + }, + { + "epoch": 1.3683926020309665, + "grad_norm": 0.659479022026062, + "learning_rate": 0.0001087764006332291, + "loss": 0.2926, + "step": 35440 + }, + { + "epoch": 1.3687787173249932, + "grad_norm": 0.9219567179679871, + "learning_rate": 0.0001087506596136273, + "loss": 0.2001, + "step": 35450 + }, + { + "epoch": 1.36916483261902, + "grad_norm": 0.8070804476737976, + "learning_rate": 0.00010872491859402551, + "loss": 0.2178, + "step": 35460 + }, + { + "epoch": 1.3695509479130468, + "grad_norm": 2.9981069564819336, + "learning_rate": 0.00010869917757442374, + "loss": 0.3079, + "step": 35470 + }, + { + "epoch": 1.3699370632070735, + "grad_norm": 0.7891242504119873, + "learning_rate": 0.00010867343655482194, + "loss": 0.2765, + "step": 35480 + }, + { + "epoch": 1.3703231785011005, + "grad_norm": 1.448637843132019, + "learning_rate": 0.00010864769553522016, + "loss": 0.3521, + "step": 35490 + }, + { + "epoch": 1.3707092937951273, + "grad_norm": 0.07628043740987778, + "learning_rate": 0.00010862195451561836, + "loss": 0.2083, + "step": 35500 + }, + { + "epoch": 1.371095409089154, + "grad_norm": 0.7549735307693481, + "learning_rate": 0.00010859621349601659, + "loss": 0.2536, + "step": 35510 + }, + { + "epoch": 1.3714815243831808, + "grad_norm": 1.3548041582107544, + "learning_rate": 0.00010857047247641479, + "loss": 0.251, + "step": 35520 + }, + { + "epoch": 1.3718676396772076, + "grad_norm": 0.530010998249054, + "learning_rate": 0.000108544731456813, + "loss": 0.1917, + "step": 35530 + }, + { + "epoch": 1.3722537549712344, + "grad_norm": 0.4148992896080017, + "learning_rate": 0.00010851899043721123, + "loss": 0.335, + "step": 35540 + }, + { + "epoch": 1.3726398702652611, + "grad_norm": 1.5118776559829712, + "learning_rate": 0.00010849324941760943, + "loss": 0.2159, + "step": 35550 + }, + { + "epoch": 1.3730259855592881, + "grad_norm": 1.036889910697937, + "learning_rate": 0.00010846750839800766, + "loss": 0.2975, + "step": 35560 + }, + { + "epoch": 1.3734121008533147, + "grad_norm": 1.724263072013855, + "learning_rate": 0.00010844176737840586, + "loss": 0.1476, + "step": 35570 + }, + { + "epoch": 1.3737982161473417, + "grad_norm": 1.599007487297058, + "learning_rate": 0.00010841602635880408, + "loss": 0.2539, + "step": 35580 + }, + { + "epoch": 1.3741843314413684, + "grad_norm": 2.9119279384613037, + "learning_rate": 0.00010839028533920228, + "loss": 0.2688, + "step": 35590 + }, + { + "epoch": 1.3745704467353952, + "grad_norm": 1.8647874593734741, + "learning_rate": 0.00010836454431960051, + "loss": 0.4158, + "step": 35600 + }, + { + "epoch": 1.374956562029422, + "grad_norm": 3.925290822982788, + "learning_rate": 0.00010833880329999872, + "loss": 0.3333, + "step": 35610 + }, + { + "epoch": 1.3753426773234487, + "grad_norm": 0.7124634385108948, + "learning_rate": 0.00010831306228039692, + "loss": 0.1069, + "step": 35620 + }, + { + "epoch": 1.3757287926174757, + "grad_norm": 1.303579330444336, + "learning_rate": 0.00010828732126079515, + "loss": 0.2898, + "step": 35630 + }, + { + "epoch": 1.3761149079115023, + "grad_norm": 3.921804189682007, + "learning_rate": 0.00010826158024119335, + "loss": 0.4212, + "step": 35640 + }, + { + "epoch": 1.3765010232055293, + "grad_norm": 1.3194564580917358, + "learning_rate": 0.00010823583922159158, + "loss": 0.2771, + "step": 35650 + }, + { + "epoch": 1.376887138499556, + "grad_norm": 1.4237637519836426, + "learning_rate": 0.00010821009820198979, + "loss": 0.2463, + "step": 35660 + }, + { + "epoch": 1.3772732537935828, + "grad_norm": 1.8165888786315918, + "learning_rate": 0.000108184357182388, + "loss": 0.291, + "step": 35670 + }, + { + "epoch": 1.3776593690876096, + "grad_norm": 1.1056426763534546, + "learning_rate": 0.00010815861616278622, + "loss": 0.2525, + "step": 35680 + }, + { + "epoch": 1.3780454843816363, + "grad_norm": 1.483189582824707, + "learning_rate": 0.00010813287514318442, + "loss": 0.1569, + "step": 35690 + }, + { + "epoch": 1.378431599675663, + "grad_norm": 1.0666841268539429, + "learning_rate": 0.00010810713412358264, + "loss": 0.235, + "step": 35700 + }, + { + "epoch": 1.3788177149696899, + "grad_norm": 1.0299845933914185, + "learning_rate": 0.00010808139310398084, + "loss": 0.3892, + "step": 35710 + }, + { + "epoch": 1.3792038302637168, + "grad_norm": 2.3474409580230713, + "learning_rate": 0.00010805565208437907, + "loss": 0.3417, + "step": 35720 + }, + { + "epoch": 1.3795899455577436, + "grad_norm": 1.7456315755844116, + "learning_rate": 0.0001080299110647773, + "loss": 0.2538, + "step": 35730 + }, + { + "epoch": 1.3799760608517704, + "grad_norm": 2.866103410720825, + "learning_rate": 0.0001080041700451755, + "loss": 0.1619, + "step": 35740 + }, + { + "epoch": 1.3803621761457971, + "grad_norm": 0.29136407375335693, + "learning_rate": 0.00010797842902557371, + "loss": 0.2692, + "step": 35750 + }, + { + "epoch": 1.380748291439824, + "grad_norm": 0.8046161532402039, + "learning_rate": 0.00010795268800597191, + "loss": 0.1575, + "step": 35760 + }, + { + "epoch": 1.3811344067338507, + "grad_norm": 0.6451787352561951, + "learning_rate": 0.00010792694698637014, + "loss": 0.4914, + "step": 35770 + }, + { + "epoch": 1.3815205220278775, + "grad_norm": 0.7289161086082458, + "learning_rate": 0.00010790120596676836, + "loss": 0.1895, + "step": 35780 + }, + { + "epoch": 1.3819066373219044, + "grad_norm": 0.8300430178642273, + "learning_rate": 0.00010787546494716656, + "loss": 0.3663, + "step": 35790 + }, + { + "epoch": 1.382292752615931, + "grad_norm": 0.17713364958763123, + "learning_rate": 0.00010784972392756479, + "loss": 0.3189, + "step": 35800 + }, + { + "epoch": 1.382678867909958, + "grad_norm": 0.903222918510437, + "learning_rate": 0.00010782398290796299, + "loss": 0.1577, + "step": 35810 + }, + { + "epoch": 1.3830649832039847, + "grad_norm": 0.08617932349443436, + "learning_rate": 0.0001077982418883612, + "loss": 0.2872, + "step": 35820 + }, + { + "epoch": 1.3834510984980115, + "grad_norm": 1.9590895175933838, + "learning_rate": 0.0001077725008687594, + "loss": 0.2907, + "step": 35830 + }, + { + "epoch": 1.3838372137920383, + "grad_norm": 1.2515161037445068, + "learning_rate": 0.00010774675984915763, + "loss": 0.177, + "step": 35840 + }, + { + "epoch": 1.384223329086065, + "grad_norm": 1.6171292066574097, + "learning_rate": 0.00010772101882955586, + "loss": 0.2321, + "step": 35850 + }, + { + "epoch": 1.3846094443800918, + "grad_norm": 0.13681405782699585, + "learning_rate": 0.00010769527780995406, + "loss": 0.247, + "step": 35860 + }, + { + "epoch": 1.3849955596741186, + "grad_norm": 1.1949968338012695, + "learning_rate": 0.00010766953679035228, + "loss": 0.2978, + "step": 35870 + }, + { + "epoch": 1.3853816749681456, + "grad_norm": 0.17001692950725555, + "learning_rate": 0.00010764379577075048, + "loss": 0.3241, + "step": 35880 + }, + { + "epoch": 1.3857677902621723, + "grad_norm": 0.8227952122688293, + "learning_rate": 0.0001076180547511487, + "loss": 0.3499, + "step": 35890 + }, + { + "epoch": 1.386153905556199, + "grad_norm": 1.4185482263565063, + "learning_rate": 0.0001075923137315469, + "loss": 0.3109, + "step": 35900 + }, + { + "epoch": 1.3865400208502259, + "grad_norm": 0.9533351063728333, + "learning_rate": 0.00010756657271194512, + "loss": 0.24, + "step": 35910 + }, + { + "epoch": 1.3869261361442526, + "grad_norm": 1.077789306640625, + "learning_rate": 0.00010754083169234335, + "loss": 0.2662, + "step": 35920 + }, + { + "epoch": 1.3873122514382794, + "grad_norm": 1.3528363704681396, + "learning_rate": 0.00010751509067274155, + "loss": 0.1623, + "step": 35930 + }, + { + "epoch": 1.3876983667323062, + "grad_norm": 0.25122806429862976, + "learning_rate": 0.00010748934965313978, + "loss": 0.1639, + "step": 35940 + }, + { + "epoch": 1.3880844820263332, + "grad_norm": 0.9446159601211548, + "learning_rate": 0.00010746360863353798, + "loss": 0.2035, + "step": 35950 + }, + { + "epoch": 1.38847059732036, + "grad_norm": 1.2258719205856323, + "learning_rate": 0.00010743786761393619, + "loss": 0.4247, + "step": 35960 + }, + { + "epoch": 1.3888567126143867, + "grad_norm": 3.117729663848877, + "learning_rate": 0.00010741212659433442, + "loss": 0.2569, + "step": 35970 + }, + { + "epoch": 1.3892428279084135, + "grad_norm": 0.6333123445510864, + "learning_rate": 0.00010738638557473262, + "loss": 0.1843, + "step": 35980 + }, + { + "epoch": 1.3896289432024402, + "grad_norm": 1.49360191822052, + "learning_rate": 0.00010736064455513084, + "loss": 0.2318, + "step": 35990 + }, + { + "epoch": 1.390015058496467, + "grad_norm": 3.9082753658294678, + "learning_rate": 0.00010733490353552904, + "loss": 0.2037, + "step": 36000 + }, + { + "epoch": 1.3904011737904938, + "grad_norm": 0.5687323808670044, + "learning_rate": 0.00010730916251592727, + "loss": 0.0871, + "step": 36010 + }, + { + "epoch": 1.3907872890845208, + "grad_norm": 0.3480868339538574, + "learning_rate": 0.00010728342149632547, + "loss": 0.4009, + "step": 36020 + }, + { + "epoch": 1.3911734043785473, + "grad_norm": 1.325042486190796, + "learning_rate": 0.00010725768047672368, + "loss": 0.5143, + "step": 36030 + }, + { + "epoch": 1.3915595196725743, + "grad_norm": 2.114786386489868, + "learning_rate": 0.00010723193945712191, + "loss": 0.2944, + "step": 36040 + }, + { + "epoch": 1.391945634966601, + "grad_norm": 1.716272234916687, + "learning_rate": 0.00010720619843752011, + "loss": 0.3793, + "step": 36050 + }, + { + "epoch": 1.3923317502606278, + "grad_norm": 2.057535171508789, + "learning_rate": 0.00010718045741791834, + "loss": 0.199, + "step": 36060 + }, + { + "epoch": 1.3927178655546546, + "grad_norm": 2.371248483657837, + "learning_rate": 0.00010715471639831653, + "loss": 0.3276, + "step": 36070 + }, + { + "epoch": 1.3931039808486814, + "grad_norm": 0.1784186065196991, + "learning_rate": 0.00010712897537871476, + "loss": 0.0922, + "step": 36080 + }, + { + "epoch": 1.3934900961427081, + "grad_norm": 0.7923040390014648, + "learning_rate": 0.00010710323435911298, + "loss": 0.1599, + "step": 36090 + }, + { + "epoch": 1.393876211436735, + "grad_norm": 1.9013831615447998, + "learning_rate": 0.00010707749333951117, + "loss": 0.4997, + "step": 36100 + }, + { + "epoch": 1.394262326730762, + "grad_norm": 1.8659415245056152, + "learning_rate": 0.0001070517523199094, + "loss": 0.1851, + "step": 36110 + }, + { + "epoch": 1.3946484420247887, + "grad_norm": 2.5775375366210938, + "learning_rate": 0.0001070260113003076, + "loss": 0.2384, + "step": 36120 + }, + { + "epoch": 1.3950345573188154, + "grad_norm": 0.21943879127502441, + "learning_rate": 0.00010700027028070583, + "loss": 0.3996, + "step": 36130 + }, + { + "epoch": 1.3954206726128422, + "grad_norm": 1.1734743118286133, + "learning_rate": 0.00010697452926110403, + "loss": 0.1737, + "step": 36140 + }, + { + "epoch": 1.395806787906869, + "grad_norm": 0.697695791721344, + "learning_rate": 0.00010694878824150225, + "loss": 0.2778, + "step": 36150 + }, + { + "epoch": 1.3961929032008957, + "grad_norm": 3.2881579399108887, + "learning_rate": 0.00010692304722190047, + "loss": 0.3198, + "step": 36160 + }, + { + "epoch": 1.3965790184949225, + "grad_norm": 0.1592467725276947, + "learning_rate": 0.00010689730620229868, + "loss": 0.2591, + "step": 36170 + }, + { + "epoch": 1.3969651337889495, + "grad_norm": 0.4579029083251953, + "learning_rate": 0.0001068715651826969, + "loss": 0.2984, + "step": 36180 + }, + { + "epoch": 1.3973512490829763, + "grad_norm": 0.4286015033721924, + "learning_rate": 0.0001068458241630951, + "loss": 0.4098, + "step": 36190 + }, + { + "epoch": 1.397737364377003, + "grad_norm": 1.7824127674102783, + "learning_rate": 0.00010682008314349332, + "loss": 0.2446, + "step": 36200 + }, + { + "epoch": 1.3981234796710298, + "grad_norm": 0.8584449887275696, + "learning_rate": 0.00010679434212389152, + "loss": 0.1415, + "step": 36210 + }, + { + "epoch": 1.3985095949650566, + "grad_norm": 1.1699339151382446, + "learning_rate": 0.00010676860110428975, + "loss": 0.1706, + "step": 36220 + }, + { + "epoch": 1.3988957102590833, + "grad_norm": 2.615877389907837, + "learning_rate": 0.00010674286008468796, + "loss": 0.2605, + "step": 36230 + }, + { + "epoch": 1.39928182555311, + "grad_norm": 2.182037591934204, + "learning_rate": 0.00010671711906508617, + "loss": 0.2067, + "step": 36240 + }, + { + "epoch": 1.399667940847137, + "grad_norm": 2.183263063430786, + "learning_rate": 0.00010669137804548439, + "loss": 0.4704, + "step": 36250 + }, + { + "epoch": 1.4000540561411636, + "grad_norm": 3.505791187286377, + "learning_rate": 0.00010666563702588259, + "loss": 0.3385, + "step": 36260 + }, + { + "epoch": 1.4004401714351906, + "grad_norm": 1.2262030839920044, + "learning_rate": 0.00010663989600628081, + "loss": 0.3999, + "step": 36270 + }, + { + "epoch": 1.4008262867292174, + "grad_norm": 2.4024577140808105, + "learning_rate": 0.00010661415498667904, + "loss": 0.1984, + "step": 36280 + }, + { + "epoch": 1.4012124020232442, + "grad_norm": 0.4166090786457062, + "learning_rate": 0.00010658841396707724, + "loss": 0.2448, + "step": 36290 + }, + { + "epoch": 1.401598517317271, + "grad_norm": 0.422590047121048, + "learning_rate": 0.00010656267294747547, + "loss": 0.2464, + "step": 36300 + }, + { + "epoch": 1.4019846326112977, + "grad_norm": 2.287503480911255, + "learning_rate": 0.00010653693192787367, + "loss": 0.1621, + "step": 36310 + }, + { + "epoch": 1.4023707479053245, + "grad_norm": 0.8126110434532166, + "learning_rate": 0.00010651119090827188, + "loss": 0.2097, + "step": 36320 + }, + { + "epoch": 1.4027568631993512, + "grad_norm": 0.683016836643219, + "learning_rate": 0.00010648544988867008, + "loss": 0.1512, + "step": 36330 + }, + { + "epoch": 1.4031429784933782, + "grad_norm": 1.5477893352508545, + "learning_rate": 0.00010645970886906831, + "loss": 0.2169, + "step": 36340 + }, + { + "epoch": 1.403529093787405, + "grad_norm": 2.183166265487671, + "learning_rate": 0.00010643396784946653, + "loss": 0.4307, + "step": 36350 + }, + { + "epoch": 1.4039152090814317, + "grad_norm": 1.5782747268676758, + "learning_rate": 0.00010640822682986473, + "loss": 0.291, + "step": 36360 + }, + { + "epoch": 1.4043013243754585, + "grad_norm": 1.2823392152786255, + "learning_rate": 0.00010638248581026296, + "loss": 0.315, + "step": 36370 + }, + { + "epoch": 1.4046874396694853, + "grad_norm": 2.0126500129699707, + "learning_rate": 0.00010635674479066116, + "loss": 0.2565, + "step": 36380 + }, + { + "epoch": 1.405073554963512, + "grad_norm": 2.490217447280884, + "learning_rate": 0.00010633100377105937, + "loss": 0.2665, + "step": 36390 + }, + { + "epoch": 1.4054596702575388, + "grad_norm": 1.2054855823516846, + "learning_rate": 0.00010630526275145757, + "loss": 0.5455, + "step": 36400 + }, + { + "epoch": 1.4058457855515658, + "grad_norm": 0.7968757748603821, + "learning_rate": 0.0001062795217318558, + "loss": 0.333, + "step": 36410 + }, + { + "epoch": 1.4062319008455926, + "grad_norm": 1.1027718782424927, + "learning_rate": 0.00010625378071225403, + "loss": 0.2019, + "step": 36420 + }, + { + "epoch": 1.4066180161396193, + "grad_norm": 1.987302541732788, + "learning_rate": 0.00010622803969265223, + "loss": 0.3159, + "step": 36430 + }, + { + "epoch": 1.407004131433646, + "grad_norm": 1.5426512956619263, + "learning_rate": 0.00010620229867305045, + "loss": 0.3759, + "step": 36440 + }, + { + "epoch": 1.4073902467276729, + "grad_norm": 0.5228156447410583, + "learning_rate": 0.00010617655765344865, + "loss": 0.2465, + "step": 36450 + }, + { + "epoch": 1.4077763620216996, + "grad_norm": 0.46890121698379517, + "learning_rate": 0.00010615081663384687, + "loss": 0.302, + "step": 36460 + }, + { + "epoch": 1.4081624773157264, + "grad_norm": 2.1506495475769043, + "learning_rate": 0.0001061250756142451, + "loss": 0.2569, + "step": 36470 + }, + { + "epoch": 1.4085485926097534, + "grad_norm": 2.307468891143799, + "learning_rate": 0.0001060993345946433, + "loss": 0.2009, + "step": 36480 + }, + { + "epoch": 1.40893470790378, + "grad_norm": 0.07033026963472366, + "learning_rate": 0.00010607359357504152, + "loss": 0.1728, + "step": 36490 + }, + { + "epoch": 1.409320823197807, + "grad_norm": 0.3262972831726074, + "learning_rate": 0.00010604785255543972, + "loss": 0.2905, + "step": 36500 + }, + { + "epoch": 1.4097069384918337, + "grad_norm": 0.755646824836731, + "learning_rate": 0.00010602211153583795, + "loss": 0.3287, + "step": 36510 + }, + { + "epoch": 1.4100930537858605, + "grad_norm": 1.1180161237716675, + "learning_rate": 0.00010599637051623615, + "loss": 0.2743, + "step": 36520 + }, + { + "epoch": 1.4104791690798872, + "grad_norm": 1.2358392477035522, + "learning_rate": 0.00010597062949663436, + "loss": 0.2672, + "step": 36530 + }, + { + "epoch": 1.410865284373914, + "grad_norm": 1.4188988208770752, + "learning_rate": 0.00010594488847703259, + "loss": 0.2552, + "step": 36540 + }, + { + "epoch": 1.4112513996679408, + "grad_norm": 1.4727978706359863, + "learning_rate": 0.00010591914745743079, + "loss": 0.2256, + "step": 36550 + }, + { + "epoch": 1.4116375149619675, + "grad_norm": 0.08973213285207748, + "learning_rate": 0.00010589340643782901, + "loss": 0.224, + "step": 36560 + }, + { + "epoch": 1.4120236302559945, + "grad_norm": 0.9915102124214172, + "learning_rate": 0.00010586766541822721, + "loss": 0.195, + "step": 36570 + }, + { + "epoch": 1.4124097455500213, + "grad_norm": 0.8524800539016724, + "learning_rate": 0.00010584192439862544, + "loss": 0.1492, + "step": 36580 + }, + { + "epoch": 1.412795860844048, + "grad_norm": 1.8414466381072998, + "learning_rate": 0.00010581618337902365, + "loss": 0.3412, + "step": 36590 + }, + { + "epoch": 1.4131819761380748, + "grad_norm": 2.596547842025757, + "learning_rate": 0.00010579044235942185, + "loss": 0.3469, + "step": 36600 + }, + { + "epoch": 1.4135680914321016, + "grad_norm": 0.5367813110351562, + "learning_rate": 0.00010576470133982008, + "loss": 0.2274, + "step": 36610 + }, + { + "epoch": 1.4139542067261284, + "grad_norm": 2.70858097076416, + "learning_rate": 0.00010573896032021828, + "loss": 0.171, + "step": 36620 + }, + { + "epoch": 1.4143403220201551, + "grad_norm": 1.7077667713165283, + "learning_rate": 0.00010571321930061651, + "loss": 0.2638, + "step": 36630 + }, + { + "epoch": 1.4147264373141821, + "grad_norm": 0.9189953804016113, + "learning_rate": 0.00010568747828101471, + "loss": 0.2283, + "step": 36640 + }, + { + "epoch": 1.4151125526082087, + "grad_norm": 2.0240087509155273, + "learning_rate": 0.00010566173726141293, + "loss": 0.3166, + "step": 36650 + }, + { + "epoch": 1.4154986679022357, + "grad_norm": 3.3304507732391357, + "learning_rate": 0.00010563599624181115, + "loss": 0.3046, + "step": 36660 + }, + { + "epoch": 1.4158847831962624, + "grad_norm": 1.955929160118103, + "learning_rate": 0.00010561025522220935, + "loss": 0.2653, + "step": 36670 + }, + { + "epoch": 1.4162708984902892, + "grad_norm": 2.2465381622314453, + "learning_rate": 0.00010558451420260757, + "loss": 0.3119, + "step": 36680 + }, + { + "epoch": 1.416657013784316, + "grad_norm": 2.108614921569824, + "learning_rate": 0.00010555877318300577, + "loss": 0.2872, + "step": 36690 + }, + { + "epoch": 1.4170431290783427, + "grad_norm": 1.35493004322052, + "learning_rate": 0.000105533032163404, + "loss": 0.1491, + "step": 36700 + }, + { + "epoch": 1.4174292443723697, + "grad_norm": 0.9102393984794617, + "learning_rate": 0.0001055072911438022, + "loss": 0.1565, + "step": 36710 + }, + { + "epoch": 1.4178153596663963, + "grad_norm": 1.7878345251083374, + "learning_rate": 0.00010548155012420043, + "loss": 0.4864, + "step": 36720 + }, + { + "epoch": 1.4182014749604233, + "grad_norm": 1.4333637952804565, + "learning_rate": 0.00010545580910459864, + "loss": 0.3246, + "step": 36730 + }, + { + "epoch": 1.41858759025445, + "grad_norm": 0.0966360941529274, + "learning_rate": 0.00010543006808499684, + "loss": 0.1758, + "step": 36740 + }, + { + "epoch": 1.4189737055484768, + "grad_norm": 0.30043545365333557, + "learning_rate": 0.00010540432706539507, + "loss": 0.1811, + "step": 36750 + }, + { + "epoch": 1.4193598208425036, + "grad_norm": 1.7705951929092407, + "learning_rate": 0.00010537858604579327, + "loss": 0.1824, + "step": 36760 + }, + { + "epoch": 1.4197459361365303, + "grad_norm": 1.2025195360183716, + "learning_rate": 0.0001053528450261915, + "loss": 0.1463, + "step": 36770 + }, + { + "epoch": 1.420132051430557, + "grad_norm": 0.3154304325580597, + "learning_rate": 0.00010532710400658972, + "loss": 0.1817, + "step": 36780 + }, + { + "epoch": 1.4205181667245839, + "grad_norm": 3.392331838607788, + "learning_rate": 0.00010530136298698792, + "loss": 0.2938, + "step": 36790 + }, + { + "epoch": 1.4209042820186109, + "grad_norm": 0.7256132364273071, + "learning_rate": 0.00010527562196738613, + "loss": 0.28, + "step": 36800 + }, + { + "epoch": 1.4212903973126376, + "grad_norm": 1.9007991552352905, + "learning_rate": 0.00010524988094778435, + "loss": 0.2306, + "step": 36810 + }, + { + "epoch": 1.4216765126066644, + "grad_norm": 3.21189546585083, + "learning_rate": 0.00010522413992818256, + "loss": 0.2945, + "step": 36820 + }, + { + "epoch": 1.4220626279006912, + "grad_norm": 0.20476600527763367, + "learning_rate": 0.00010519839890858076, + "loss": 0.3129, + "step": 36830 + }, + { + "epoch": 1.422448743194718, + "grad_norm": 1.4225107431411743, + "learning_rate": 0.00010517265788897899, + "loss": 0.2262, + "step": 36840 + }, + { + "epoch": 1.4228348584887447, + "grad_norm": 1.203728437423706, + "learning_rate": 0.00010514691686937721, + "loss": 0.176, + "step": 36850 + }, + { + "epoch": 1.4232209737827715, + "grad_norm": 0.3001759648323059, + "learning_rate": 0.00010512117584977541, + "loss": 0.218, + "step": 36860 + }, + { + "epoch": 1.4236070890767984, + "grad_norm": 2.3143389225006104, + "learning_rate": 0.00010509543483017363, + "loss": 0.1562, + "step": 36870 + }, + { + "epoch": 1.423993204370825, + "grad_norm": 0.5087364912033081, + "learning_rate": 0.00010506969381057184, + "loss": 0.1207, + "step": 36880 + }, + { + "epoch": 1.424379319664852, + "grad_norm": 1.6521960496902466, + "learning_rate": 0.00010504395279097005, + "loss": 0.4756, + "step": 36890 + }, + { + "epoch": 1.4247654349588788, + "grad_norm": 0.09236706793308258, + "learning_rate": 0.00010501821177136825, + "loss": 0.1755, + "step": 36900 + }, + { + "epoch": 1.4251515502529055, + "grad_norm": 0.2143094390630722, + "learning_rate": 0.00010499247075176648, + "loss": 0.2126, + "step": 36910 + }, + { + "epoch": 1.4255376655469323, + "grad_norm": 1.210170865058899, + "learning_rate": 0.00010496672973216471, + "loss": 0.1391, + "step": 36920 + }, + { + "epoch": 1.425923780840959, + "grad_norm": 0.6248244047164917, + "learning_rate": 0.00010494098871256291, + "loss": 0.2898, + "step": 36930 + }, + { + "epoch": 1.426309896134986, + "grad_norm": 2.1100337505340576, + "learning_rate": 0.00010491524769296113, + "loss": 0.1198, + "step": 36940 + }, + { + "epoch": 1.4266960114290126, + "grad_norm": 2.5673348903656006, + "learning_rate": 0.00010488950667335933, + "loss": 0.2284, + "step": 36950 + }, + { + "epoch": 1.4270821267230396, + "grad_norm": 1.868195652961731, + "learning_rate": 0.00010486376565375755, + "loss": 0.3738, + "step": 36960 + }, + { + "epoch": 1.4274682420170663, + "grad_norm": 1.0951671600341797, + "learning_rate": 0.00010483802463415577, + "loss": 0.1424, + "step": 36970 + }, + { + "epoch": 1.4278543573110931, + "grad_norm": 0.4791143536567688, + "learning_rate": 0.00010481228361455397, + "loss": 0.1445, + "step": 36980 + }, + { + "epoch": 1.4282404726051199, + "grad_norm": 3.2477540969848633, + "learning_rate": 0.0001047865425949522, + "loss": 0.2676, + "step": 36990 + }, + { + "epoch": 1.4286265878991466, + "grad_norm": 0.8082342147827148, + "learning_rate": 0.0001047608015753504, + "loss": 0.2295, + "step": 37000 + }, + { + "epoch": 1.4290127031931734, + "grad_norm": 2.3048954010009766, + "learning_rate": 0.00010473506055574863, + "loss": 0.1996, + "step": 37010 + }, + { + "epoch": 1.4293988184872002, + "grad_norm": 0.42648783326148987, + "learning_rate": 0.00010470931953614683, + "loss": 0.2549, + "step": 37020 + }, + { + "epoch": 1.4297849337812272, + "grad_norm": 1.8794362545013428, + "learning_rate": 0.00010468357851654504, + "loss": 0.1929, + "step": 37030 + }, + { + "epoch": 1.430171049075254, + "grad_norm": 0.9679039120674133, + "learning_rate": 0.00010465783749694327, + "loss": 0.1544, + "step": 37040 + }, + { + "epoch": 1.4305571643692807, + "grad_norm": 0.7789367437362671, + "learning_rate": 0.00010463209647734147, + "loss": 0.2251, + "step": 37050 + }, + { + "epoch": 1.4309432796633075, + "grad_norm": 1.1705437898635864, + "learning_rate": 0.0001046063554577397, + "loss": 0.1567, + "step": 37060 + }, + { + "epoch": 1.4313293949573342, + "grad_norm": 0.13087430596351624, + "learning_rate": 0.00010458061443813789, + "loss": 0.3237, + "step": 37070 + }, + { + "epoch": 1.431715510251361, + "grad_norm": 1.0658761262893677, + "learning_rate": 0.00010455487341853612, + "loss": 0.24, + "step": 37080 + }, + { + "epoch": 1.4321016255453878, + "grad_norm": 1.686922311782837, + "learning_rate": 0.00010452913239893433, + "loss": 0.1911, + "step": 37090 + }, + { + "epoch": 1.4324877408394148, + "grad_norm": 2.010221481323242, + "learning_rate": 0.00010450339137933253, + "loss": 0.4569, + "step": 37100 + }, + { + "epoch": 1.4328738561334413, + "grad_norm": 0.8007562160491943, + "learning_rate": 0.00010447765035973076, + "loss": 0.1198, + "step": 37110 + }, + { + "epoch": 1.4332599714274683, + "grad_norm": 0.5455211997032166, + "learning_rate": 0.00010445190934012896, + "loss": 0.2179, + "step": 37120 + }, + { + "epoch": 1.433646086721495, + "grad_norm": 1.4133542776107788, + "learning_rate": 0.00010442616832052719, + "loss": 0.45, + "step": 37130 + }, + { + "epoch": 1.4340322020155218, + "grad_norm": 1.6218222379684448, + "learning_rate": 0.00010440042730092539, + "loss": 0.19, + "step": 37140 + }, + { + "epoch": 1.4344183173095486, + "grad_norm": 0.6425970196723938, + "learning_rate": 0.00010437468628132361, + "loss": 0.2345, + "step": 37150 + }, + { + "epoch": 1.4348044326035754, + "grad_norm": 0.4344918131828308, + "learning_rate": 0.00010434894526172183, + "loss": 0.2532, + "step": 37160 + }, + { + "epoch": 1.4351905478976021, + "grad_norm": 0.6628998517990112, + "learning_rate": 0.00010432320424212003, + "loss": 0.2281, + "step": 37170 + }, + { + "epoch": 1.435576663191629, + "grad_norm": 0.8695842027664185, + "learning_rate": 0.00010429746322251825, + "loss": 0.2796, + "step": 37180 + }, + { + "epoch": 1.435962778485656, + "grad_norm": 0.16481854021549225, + "learning_rate": 0.00010427172220291645, + "loss": 0.2333, + "step": 37190 + }, + { + "epoch": 1.4363488937796827, + "grad_norm": 0.7194454073905945, + "learning_rate": 0.00010424598118331468, + "loss": 0.1413, + "step": 37200 + }, + { + "epoch": 1.4367350090737094, + "grad_norm": 4.845508575439453, + "learning_rate": 0.00010422024016371288, + "loss": 0.3944, + "step": 37210 + }, + { + "epoch": 1.4371211243677362, + "grad_norm": 2.6386618614196777, + "learning_rate": 0.0001041944991441111, + "loss": 0.3037, + "step": 37220 + }, + { + "epoch": 1.437507239661763, + "grad_norm": 0.4089922308921814, + "learning_rate": 0.00010416875812450932, + "loss": 0.2985, + "step": 37230 + }, + { + "epoch": 1.4378933549557897, + "grad_norm": 1.456944465637207, + "learning_rate": 0.00010414301710490752, + "loss": 0.3031, + "step": 37240 + }, + { + "epoch": 1.4382794702498165, + "grad_norm": 1.301829218864441, + "learning_rate": 0.00010411727608530575, + "loss": 0.2578, + "step": 37250 + }, + { + "epoch": 1.4386655855438435, + "grad_norm": 1.2072703838348389, + "learning_rate": 0.00010409153506570395, + "loss": 0.391, + "step": 37260 + }, + { + "epoch": 1.4390517008378703, + "grad_norm": 0.5538531541824341, + "learning_rate": 0.00010406579404610217, + "loss": 0.1867, + "step": 37270 + }, + { + "epoch": 1.439437816131897, + "grad_norm": 1.0898678302764893, + "learning_rate": 0.0001040400530265004, + "loss": 0.2112, + "step": 37280 + }, + { + "epoch": 1.4398239314259238, + "grad_norm": 1.5769239664077759, + "learning_rate": 0.0001040143120068986, + "loss": 0.3121, + "step": 37290 + }, + { + "epoch": 1.4402100467199506, + "grad_norm": 0.3964422345161438, + "learning_rate": 0.00010398857098729681, + "loss": 0.196, + "step": 37300 + }, + { + "epoch": 1.4405961620139773, + "grad_norm": 1.0268182754516602, + "learning_rate": 0.00010396282996769501, + "loss": 0.1785, + "step": 37310 + }, + { + "epoch": 1.440982277308004, + "grad_norm": 1.750826358795166, + "learning_rate": 0.00010393708894809324, + "loss": 0.3654, + "step": 37320 + }, + { + "epoch": 1.441368392602031, + "grad_norm": 1.1231745481491089, + "learning_rate": 0.00010391134792849144, + "loss": 0.2594, + "step": 37330 + }, + { + "epoch": 1.4417545078960576, + "grad_norm": 0.2897786498069763, + "learning_rate": 0.00010388560690888967, + "loss": 0.3483, + "step": 37340 + }, + { + "epoch": 1.4421406231900846, + "grad_norm": 0.07170752435922623, + "learning_rate": 0.00010385986588928789, + "loss": 0.1462, + "step": 37350 + }, + { + "epoch": 1.4425267384841114, + "grad_norm": 3.326099395751953, + "learning_rate": 0.00010383412486968609, + "loss": 0.3785, + "step": 37360 + }, + { + "epoch": 1.4429128537781382, + "grad_norm": 0.5063263773918152, + "learning_rate": 0.0001038083838500843, + "loss": 0.2974, + "step": 37370 + }, + { + "epoch": 1.443298969072165, + "grad_norm": 1.160088062286377, + "learning_rate": 0.00010378264283048252, + "loss": 0.3003, + "step": 37380 + }, + { + "epoch": 1.4436850843661917, + "grad_norm": 0.7043284177780151, + "learning_rate": 0.00010375690181088073, + "loss": 0.3193, + "step": 37390 + }, + { + "epoch": 1.4440711996602185, + "grad_norm": 0.2916620373725891, + "learning_rate": 0.00010373116079127896, + "loss": 0.235, + "step": 37400 + }, + { + "epoch": 1.4444573149542452, + "grad_norm": 2.1940219402313232, + "learning_rate": 0.00010370541977167716, + "loss": 0.2657, + "step": 37410 + }, + { + "epoch": 1.4448434302482722, + "grad_norm": 1.849794626235962, + "learning_rate": 0.00010367967875207539, + "loss": 0.3121, + "step": 37420 + }, + { + "epoch": 1.445229545542299, + "grad_norm": 1.1139589548110962, + "learning_rate": 0.00010365393773247359, + "loss": 0.2179, + "step": 37430 + }, + { + "epoch": 1.4456156608363258, + "grad_norm": 0.917142927646637, + "learning_rate": 0.0001036281967128718, + "loss": 0.2406, + "step": 37440 + }, + { + "epoch": 1.4460017761303525, + "grad_norm": 1.6973673105239868, + "learning_rate": 0.00010360245569327001, + "loss": 0.3236, + "step": 37450 + }, + { + "epoch": 1.4463878914243793, + "grad_norm": 0.3979933559894562, + "learning_rate": 0.00010357671467366823, + "loss": 0.1477, + "step": 37460 + }, + { + "epoch": 1.446774006718406, + "grad_norm": 0.8938451409339905, + "learning_rate": 0.00010355097365406645, + "loss": 0.116, + "step": 37470 + }, + { + "epoch": 1.4471601220124328, + "grad_norm": 1.068787693977356, + "learning_rate": 0.00010352523263446465, + "loss": 0.2342, + "step": 37480 + }, + { + "epoch": 1.4475462373064598, + "grad_norm": 1.1985591650009155, + "learning_rate": 0.00010349949161486288, + "loss": 0.2423, + "step": 37490 + }, + { + "epoch": 1.4479323526004866, + "grad_norm": 1.0383973121643066, + "learning_rate": 0.00010347375059526108, + "loss": 0.3059, + "step": 37500 + }, + { + "epoch": 1.4483184678945134, + "grad_norm": 0.26652297377586365, + "learning_rate": 0.0001034480095756593, + "loss": 0.262, + "step": 37510 + }, + { + "epoch": 1.4487045831885401, + "grad_norm": 1.8498083353042603, + "learning_rate": 0.0001034222685560575, + "loss": 0.2122, + "step": 37520 + }, + { + "epoch": 1.4490906984825669, + "grad_norm": 1.2896068096160889, + "learning_rate": 0.00010339652753645572, + "loss": 0.2939, + "step": 37530 + }, + { + "epoch": 1.4494768137765937, + "grad_norm": 1.423343300819397, + "learning_rate": 0.00010337078651685395, + "loss": 0.3375, + "step": 37540 + }, + { + "epoch": 1.4498629290706204, + "grad_norm": 1.8248246908187866, + "learning_rate": 0.00010334504549725215, + "loss": 0.3534, + "step": 37550 + }, + { + "epoch": 1.4502490443646474, + "grad_norm": 2.3713393211364746, + "learning_rate": 0.00010331930447765037, + "loss": 0.2795, + "step": 37560 + }, + { + "epoch": 1.450635159658674, + "grad_norm": 2.1431849002838135, + "learning_rate": 0.00010329356345804857, + "loss": 0.2503, + "step": 37570 + }, + { + "epoch": 1.451021274952701, + "grad_norm": 1.6521297693252563, + "learning_rate": 0.0001032678224384468, + "loss": 0.1789, + "step": 37580 + }, + { + "epoch": 1.4514073902467277, + "grad_norm": 1.5589754581451416, + "learning_rate": 0.00010324208141884501, + "loss": 0.3201, + "step": 37590 + }, + { + "epoch": 1.4517935055407545, + "grad_norm": 0.4339803159236908, + "learning_rate": 0.00010321634039924321, + "loss": 0.2964, + "step": 37600 + }, + { + "epoch": 1.4521796208347812, + "grad_norm": 0.30054792761802673, + "learning_rate": 0.00010319059937964144, + "loss": 0.3477, + "step": 37610 + }, + { + "epoch": 1.452565736128808, + "grad_norm": 2.2864038944244385, + "learning_rate": 0.00010316485836003964, + "loss": 0.2849, + "step": 37620 + }, + { + "epoch": 1.4529518514228348, + "grad_norm": 1.8392651081085205, + "learning_rate": 0.00010313911734043787, + "loss": 0.2089, + "step": 37630 + }, + { + "epoch": 1.4533379667168616, + "grad_norm": 1.0444347858428955, + "learning_rate": 0.00010311337632083607, + "loss": 0.2306, + "step": 37640 + }, + { + "epoch": 1.4537240820108885, + "grad_norm": 1.265647053718567, + "learning_rate": 0.00010308763530123429, + "loss": 0.2192, + "step": 37650 + }, + { + "epoch": 1.4541101973049153, + "grad_norm": 1.0464913845062256, + "learning_rate": 0.0001030618942816325, + "loss": 0.2857, + "step": 37660 + }, + { + "epoch": 1.454496312598942, + "grad_norm": 0.5931240320205688, + "learning_rate": 0.0001030361532620307, + "loss": 0.1751, + "step": 37670 + }, + { + "epoch": 1.4548824278929688, + "grad_norm": 2.0155346393585205, + "learning_rate": 0.00010301041224242893, + "loss": 0.4124, + "step": 37680 + }, + { + "epoch": 1.4552685431869956, + "grad_norm": 2.843345880508423, + "learning_rate": 0.00010298467122282713, + "loss": 0.2119, + "step": 37690 + }, + { + "epoch": 1.4556546584810224, + "grad_norm": 1.977612018585205, + "learning_rate": 0.00010295893020322536, + "loss": 0.271, + "step": 37700 + }, + { + "epoch": 1.4560407737750491, + "grad_norm": 1.2862237691879272, + "learning_rate": 0.00010293318918362356, + "loss": 0.4147, + "step": 37710 + }, + { + "epoch": 1.4564268890690761, + "grad_norm": 0.8235340118408203, + "learning_rate": 0.00010290744816402179, + "loss": 0.1918, + "step": 37720 + }, + { + "epoch": 1.456813004363103, + "grad_norm": 1.0351753234863281, + "learning_rate": 0.00010288170714442, + "loss": 0.2343, + "step": 37730 + }, + { + "epoch": 1.4571991196571297, + "grad_norm": 0.2567160725593567, + "learning_rate": 0.0001028559661248182, + "loss": 0.2664, + "step": 37740 + }, + { + "epoch": 1.4575852349511564, + "grad_norm": 2.7498669624328613, + "learning_rate": 0.00010283022510521643, + "loss": 0.2839, + "step": 37750 + }, + { + "epoch": 1.4579713502451832, + "grad_norm": 0.7125422358512878, + "learning_rate": 0.00010280448408561463, + "loss": 0.2411, + "step": 37760 + }, + { + "epoch": 1.45835746553921, + "grad_norm": 0.720761239528656, + "learning_rate": 0.00010277874306601285, + "loss": 0.315, + "step": 37770 + }, + { + "epoch": 1.4587435808332367, + "grad_norm": 3.188563346862793, + "learning_rate": 0.00010275300204641108, + "loss": 0.3945, + "step": 37780 + }, + { + "epoch": 1.4591296961272637, + "grad_norm": 1.23422110080719, + "learning_rate": 0.00010272726102680928, + "loss": 0.1828, + "step": 37790 + }, + { + "epoch": 1.4595158114212903, + "grad_norm": 1.1572456359863281, + "learning_rate": 0.00010270152000720749, + "loss": 0.1687, + "step": 37800 + }, + { + "epoch": 1.4599019267153173, + "grad_norm": 0.6565262079238892, + "learning_rate": 0.00010267577898760569, + "loss": 0.3145, + "step": 37810 + }, + { + "epoch": 1.460288042009344, + "grad_norm": 0.6239646673202515, + "learning_rate": 0.00010265003796800392, + "loss": 0.2381, + "step": 37820 + }, + { + "epoch": 1.4606741573033708, + "grad_norm": 0.7578912377357483, + "learning_rate": 0.00010262429694840212, + "loss": 0.2019, + "step": 37830 + }, + { + "epoch": 1.4610602725973976, + "grad_norm": 1.0978549718856812, + "learning_rate": 0.00010259855592880035, + "loss": 0.2477, + "step": 37840 + }, + { + "epoch": 1.4614463878914243, + "grad_norm": 0.3615519404411316, + "learning_rate": 0.00010257281490919857, + "loss": 0.4241, + "step": 37850 + }, + { + "epoch": 1.461832503185451, + "grad_norm": 0.10355047881603241, + "learning_rate": 0.00010254707388959677, + "loss": 0.1181, + "step": 37860 + }, + { + "epoch": 1.4622186184794779, + "grad_norm": 0.23957425355911255, + "learning_rate": 0.00010252133286999499, + "loss": 0.1675, + "step": 37870 + }, + { + "epoch": 1.4626047337735049, + "grad_norm": 1.7581062316894531, + "learning_rate": 0.00010249559185039318, + "loss": 0.241, + "step": 37880 + }, + { + "epoch": 1.4629908490675316, + "grad_norm": 1.660989761352539, + "learning_rate": 0.00010246985083079141, + "loss": 0.2099, + "step": 37890 + }, + { + "epoch": 1.4633769643615584, + "grad_norm": 1.1359142065048218, + "learning_rate": 0.00010244410981118964, + "loss": 0.3686, + "step": 37900 + }, + { + "epoch": 1.4637630796555852, + "grad_norm": 0.9489149451255798, + "learning_rate": 0.00010241836879158784, + "loss": 0.3616, + "step": 37910 + }, + { + "epoch": 1.464149194949612, + "grad_norm": 1.5209956169128418, + "learning_rate": 0.00010239262777198607, + "loss": 0.3159, + "step": 37920 + }, + { + "epoch": 1.4645353102436387, + "grad_norm": 0.848943293094635, + "learning_rate": 0.00010236688675238427, + "loss": 0.3228, + "step": 37930 + }, + { + "epoch": 1.4649214255376655, + "grad_norm": 1.1476777791976929, + "learning_rate": 0.00010234114573278248, + "loss": 0.1873, + "step": 37940 + }, + { + "epoch": 1.4653075408316925, + "grad_norm": 1.4093862771987915, + "learning_rate": 0.00010231540471318068, + "loss": 0.5554, + "step": 37950 + }, + { + "epoch": 1.465693656125719, + "grad_norm": 0.17349161207675934, + "learning_rate": 0.0001022896636935789, + "loss": 0.2992, + "step": 37960 + }, + { + "epoch": 1.466079771419746, + "grad_norm": 0.28885993361473083, + "learning_rate": 0.00010226392267397713, + "loss": 0.2618, + "step": 37970 + }, + { + "epoch": 1.4664658867137728, + "grad_norm": 1.1087830066680908, + "learning_rate": 0.00010223818165437533, + "loss": 0.2647, + "step": 37980 + }, + { + "epoch": 1.4668520020077995, + "grad_norm": 0.08338876068592072, + "learning_rate": 0.00010221244063477356, + "loss": 0.3057, + "step": 37990 + }, + { + "epoch": 1.4672381173018263, + "grad_norm": 2.159362316131592, + "learning_rate": 0.00010218669961517176, + "loss": 0.1499, + "step": 38000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.314039984128e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2cab5c3bd4d34bc664eaedc5529cd2532162a985 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:444f72342c85df25345ce98d4bf2fcb3c5554a3c3997cc2d4060116579b9c946 +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..aaaa09f132cf25750d2a6b4c6c500c5de144a838 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0770a639dedba7c5724cf8fed785fef70ee5e88c9cd60410f55a0a1820899130 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..137dcaf3a2256e078a6f255754f861411fa15e8d Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..7b1f42efdf11ddd236e3c43a8479a7428c9797b0 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..90ff2a2c2f1e0c053b0c00e269f1f948b80963f7 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/trainer_state.json @@ -0,0 +1,26984 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.4865438820031662, + "eval_steps": 500, + "global_step": 38500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + }, + { + "epoch": 1.274566585582455, + "grad_norm": 0.5019121766090393, + "learning_rate": 0.0001150314683964632, + "loss": 0.1857, + "step": 33010 + }, + { + "epoch": 1.2749527008764816, + "grad_norm": 0.258952260017395, + "learning_rate": 0.0001150057273768614, + "loss": 0.11, + "step": 33020 + }, + { + "epoch": 1.2753388161705086, + "grad_norm": 0.8540348410606384, + "learning_rate": 0.00011497998635725962, + "loss": 0.1852, + "step": 33030 + }, + { + "epoch": 1.2757249314645354, + "grad_norm": 0.08689398318529129, + "learning_rate": 0.00011495424533765783, + "loss": 0.2225, + "step": 33040 + }, + { + "epoch": 1.2761110467585621, + "grad_norm": 0.42253822088241577, + "learning_rate": 0.00011492850431805605, + "loss": 0.0751, + "step": 33050 + }, + { + "epoch": 1.276497162052589, + "grad_norm": 1.2964017391204834, + "learning_rate": 0.00011490276329845426, + "loss": 0.2384, + "step": 33060 + }, + { + "epoch": 1.2768832773466157, + "grad_norm": 0.5337836146354675, + "learning_rate": 0.00011487702227885246, + "loss": 0.1415, + "step": 33070 + }, + { + "epoch": 1.2772693926406424, + "grad_norm": 2.7771682739257812, + "learning_rate": 0.00011485128125925069, + "loss": 0.378, + "step": 33080 + }, + { + "epoch": 1.2776555079346692, + "grad_norm": 1.5107232332229614, + "learning_rate": 0.00011482554023964889, + "loss": 0.2482, + "step": 33090 + }, + { + "epoch": 1.2780416232286962, + "grad_norm": 0.6885499358177185, + "learning_rate": 0.00011479979922004711, + "loss": 0.2122, + "step": 33100 + }, + { + "epoch": 1.278427738522723, + "grad_norm": 0.9016557335853577, + "learning_rate": 0.00011477405820044533, + "loss": 0.2841, + "step": 33110 + }, + { + "epoch": 1.2788138538167497, + "grad_norm": 1.9532525539398193, + "learning_rate": 0.00011474831718084354, + "loss": 0.2281, + "step": 33120 + }, + { + "epoch": 1.2791999691107765, + "grad_norm": 2.1078782081604004, + "learning_rate": 0.00011472257616124175, + "loss": 0.2836, + "step": 33130 + }, + { + "epoch": 1.2795860844048033, + "grad_norm": 0.19830390810966492, + "learning_rate": 0.00011469683514163995, + "loss": 0.388, + "step": 33140 + }, + { + "epoch": 1.27997219969883, + "grad_norm": 0.17538850009441376, + "learning_rate": 0.00011467109412203818, + "loss": 0.3274, + "step": 33150 + }, + { + "epoch": 1.2803583149928568, + "grad_norm": 0.7402139902114868, + "learning_rate": 0.00011464535310243641, + "loss": 0.1979, + "step": 33160 + }, + { + "epoch": 1.2807444302868838, + "grad_norm": 0.2097146362066269, + "learning_rate": 0.00011461961208283461, + "loss": 0.2464, + "step": 33170 + }, + { + "epoch": 1.2811305455809103, + "grad_norm": 1.2441083192825317, + "learning_rate": 0.00011459387106323283, + "loss": 0.266, + "step": 33180 + }, + { + "epoch": 1.2815166608749373, + "grad_norm": 2.518852710723877, + "learning_rate": 0.00011456813004363103, + "loss": 0.253, + "step": 33190 + }, + { + "epoch": 1.281902776168964, + "grad_norm": 0.8078998327255249, + "learning_rate": 0.00011454238902402925, + "loss": 0.2361, + "step": 33200 + }, + { + "epoch": 1.2822888914629909, + "grad_norm": 1.2297371625900269, + "learning_rate": 0.00011451664800442745, + "loss": 0.1974, + "step": 33210 + }, + { + "epoch": 1.2826750067570176, + "grad_norm": 0.4303855895996094, + "learning_rate": 0.00011449090698482567, + "loss": 0.3563, + "step": 33220 + }, + { + "epoch": 1.2830611220510444, + "grad_norm": 1.3215210437774658, + "learning_rate": 0.0001144651659652239, + "loss": 0.2818, + "step": 33230 + }, + { + "epoch": 1.2834472373450712, + "grad_norm": 1.546265959739685, + "learning_rate": 0.0001144394249456221, + "loss": 0.5778, + "step": 33240 + }, + { + "epoch": 1.283833352639098, + "grad_norm": 0.8895953297615051, + "learning_rate": 0.00011441368392602033, + "loss": 0.2485, + "step": 33250 + }, + { + "epoch": 1.284219467933125, + "grad_norm": 0.7534870505332947, + "learning_rate": 0.00011438794290641853, + "loss": 0.2825, + "step": 33260 + }, + { + "epoch": 1.2846055832271517, + "grad_norm": 0.052820973098278046, + "learning_rate": 0.00011436220188681674, + "loss": 0.2191, + "step": 33270 + }, + { + "epoch": 1.2849916985211784, + "grad_norm": 0.9264475107192993, + "learning_rate": 0.00011433646086721494, + "loss": 0.181, + "step": 33280 + }, + { + "epoch": 1.2853778138152052, + "grad_norm": 0.2128441333770752, + "learning_rate": 0.00011431071984761317, + "loss": 0.1819, + "step": 33290 + }, + { + "epoch": 1.285763929109232, + "grad_norm": 0.5400950312614441, + "learning_rate": 0.0001142849788280114, + "loss": 0.4316, + "step": 33300 + }, + { + "epoch": 1.2861500444032588, + "grad_norm": 0.8033271431922913, + "learning_rate": 0.00011425923780840959, + "loss": 0.2146, + "step": 33310 + }, + { + "epoch": 1.2865361596972855, + "grad_norm": 2.012575149536133, + "learning_rate": 0.00011423349678880782, + "loss": 0.4335, + "step": 33320 + }, + { + "epoch": 1.2869222749913125, + "grad_norm": 0.7352376580238342, + "learning_rate": 0.00011420775576920602, + "loss": 0.2124, + "step": 33330 + }, + { + "epoch": 1.2873083902853393, + "grad_norm": 0.769036591053009, + "learning_rate": 0.00011418201474960423, + "loss": 0.3602, + "step": 33340 + }, + { + "epoch": 1.287694505579366, + "grad_norm": 0.250592976808548, + "learning_rate": 0.00011415627373000246, + "loss": 0.1692, + "step": 33350 + }, + { + "epoch": 1.2880806208733928, + "grad_norm": 2.43820858001709, + "learning_rate": 0.00011413053271040066, + "loss": 0.2777, + "step": 33360 + }, + { + "epoch": 1.2884667361674196, + "grad_norm": 1.3179954290390015, + "learning_rate": 0.00011410479169079889, + "loss": 0.1794, + "step": 33370 + }, + { + "epoch": 1.2888528514614463, + "grad_norm": 1.0040466785430908, + "learning_rate": 0.00011407905067119709, + "loss": 0.3037, + "step": 33380 + }, + { + "epoch": 1.2892389667554731, + "grad_norm": 5.296288013458252, + "learning_rate": 0.00011405330965159531, + "loss": 0.2904, + "step": 33390 + }, + { + "epoch": 1.2896250820495, + "grad_norm": 0.4267273247241974, + "learning_rate": 0.00011402756863199351, + "loss": 0.2263, + "step": 33400 + }, + { + "epoch": 1.2900111973435266, + "grad_norm": 0.8817713260650635, + "learning_rate": 0.00011400182761239173, + "loss": 0.2715, + "step": 33410 + }, + { + "epoch": 1.2903973126375536, + "grad_norm": 2.7891275882720947, + "learning_rate": 0.00011397608659278995, + "loss": 0.2781, + "step": 33420 + }, + { + "epoch": 1.2907834279315804, + "grad_norm": 0.3013952374458313, + "learning_rate": 0.00011395034557318815, + "loss": 0.2563, + "step": 33430 + }, + { + "epoch": 1.2911695432256072, + "grad_norm": 1.766413927078247, + "learning_rate": 0.00011392460455358638, + "loss": 0.1854, + "step": 33440 + }, + { + "epoch": 1.291555658519634, + "grad_norm": 0.25331103801727295, + "learning_rate": 0.00011389886353398458, + "loss": 0.1206, + "step": 33450 + }, + { + "epoch": 1.2919417738136607, + "grad_norm": 0.036400288343429565, + "learning_rate": 0.0001138731225143828, + "loss": 0.4707, + "step": 33460 + }, + { + "epoch": 1.2923278891076875, + "grad_norm": 1.5620888471603394, + "learning_rate": 0.00011384738149478102, + "loss": 0.3312, + "step": 33470 + }, + { + "epoch": 1.2927140044017142, + "grad_norm": 0.6670392155647278, + "learning_rate": 0.00011382164047517922, + "loss": 0.2341, + "step": 33480 + }, + { + "epoch": 1.2931001196957412, + "grad_norm": 2.3108737468719482, + "learning_rate": 0.00011379589945557745, + "loss": 0.3843, + "step": 33490 + }, + { + "epoch": 1.293486234989768, + "grad_norm": 0.8025147318840027, + "learning_rate": 0.00011377015843597565, + "loss": 0.1982, + "step": 33500 + }, + { + "epoch": 1.2938723502837948, + "grad_norm": 1.7835719585418701, + "learning_rate": 0.00011374441741637387, + "loss": 0.3285, + "step": 33510 + }, + { + "epoch": 1.2942584655778215, + "grad_norm": 2.041508913040161, + "learning_rate": 0.00011371867639677207, + "loss": 0.2044, + "step": 33520 + }, + { + "epoch": 1.2946445808718483, + "grad_norm": 1.103378415107727, + "learning_rate": 0.0001136929353771703, + "loss": 0.1682, + "step": 33530 + }, + { + "epoch": 1.295030696165875, + "grad_norm": 0.057376351207494736, + "learning_rate": 0.00011366719435756851, + "loss": 0.1642, + "step": 33540 + }, + { + "epoch": 1.2954168114599018, + "grad_norm": 0.6539410948753357, + "learning_rate": 0.00011364145333796671, + "loss": 0.1549, + "step": 33550 + }, + { + "epoch": 1.2958029267539288, + "grad_norm": 1.250543236732483, + "learning_rate": 0.00011361571231836494, + "loss": 0.3764, + "step": 33560 + }, + { + "epoch": 1.2961890420479556, + "grad_norm": 0.23697887361049652, + "learning_rate": 0.00011358997129876314, + "loss": 0.3999, + "step": 33570 + }, + { + "epoch": 1.2965751573419824, + "grad_norm": 0.9318505525588989, + "learning_rate": 0.00011356423027916137, + "loss": 0.4156, + "step": 33580 + }, + { + "epoch": 1.2969612726360091, + "grad_norm": 1.3910777568817139, + "learning_rate": 0.00011353848925955957, + "loss": 0.3455, + "step": 33590 + }, + { + "epoch": 1.297347387930036, + "grad_norm": 1.6764451265335083, + "learning_rate": 0.00011351274823995779, + "loss": 0.1884, + "step": 33600 + }, + { + "epoch": 1.2977335032240627, + "grad_norm": 0.9300051927566528, + "learning_rate": 0.000113487007220356, + "loss": 0.119, + "step": 33610 + }, + { + "epoch": 1.2981196185180894, + "grad_norm": 2.447462558746338, + "learning_rate": 0.00011346126620075422, + "loss": 0.4403, + "step": 33620 + }, + { + "epoch": 1.2985057338121164, + "grad_norm": 1.216407060623169, + "learning_rate": 0.00011343552518115243, + "loss": 0.2415, + "step": 33630 + }, + { + "epoch": 1.298891849106143, + "grad_norm": 2.968648910522461, + "learning_rate": 0.00011340978416155063, + "loss": 0.2899, + "step": 33640 + }, + { + "epoch": 1.29927796440017, + "grad_norm": 0.6649970412254333, + "learning_rate": 0.00011338404314194886, + "loss": 0.3809, + "step": 33650 + }, + { + "epoch": 1.2996640796941967, + "grad_norm": 1.7277917861938477, + "learning_rate": 0.00011335830212234709, + "loss": 0.3308, + "step": 33660 + }, + { + "epoch": 1.3000501949882235, + "grad_norm": 1.3269709348678589, + "learning_rate": 0.00011333256110274529, + "loss": 0.3682, + "step": 33670 + }, + { + "epoch": 1.3004363102822503, + "grad_norm": 0.20609407126903534, + "learning_rate": 0.0001133068200831435, + "loss": 0.1379, + "step": 33680 + }, + { + "epoch": 1.300822425576277, + "grad_norm": 0.6592215299606323, + "learning_rate": 0.00011328107906354171, + "loss": 0.2746, + "step": 33690 + }, + { + "epoch": 1.3012085408703038, + "grad_norm": 1.903635859489441, + "learning_rate": 0.00011325533804393993, + "loss": 0.4729, + "step": 33700 + }, + { + "epoch": 1.3015946561643306, + "grad_norm": 0.8432504534721375, + "learning_rate": 0.00011322959702433813, + "loss": 0.2835, + "step": 33710 + }, + { + "epoch": 1.3019807714583576, + "grad_norm": 0.9862542152404785, + "learning_rate": 0.00011320385600473635, + "loss": 0.1823, + "step": 33720 + }, + { + "epoch": 1.3023668867523843, + "grad_norm": 3.845738649368286, + "learning_rate": 0.00011317811498513458, + "loss": 0.2542, + "step": 33730 + }, + { + "epoch": 1.302753002046411, + "grad_norm": 0.6317747235298157, + "learning_rate": 0.00011315237396553278, + "loss": 0.22, + "step": 33740 + }, + { + "epoch": 1.3031391173404379, + "grad_norm": 2.5221354961395264, + "learning_rate": 0.000113126632945931, + "loss": 0.2253, + "step": 33750 + }, + { + "epoch": 1.3035252326344646, + "grad_norm": 1.3326247930526733, + "learning_rate": 0.0001131008919263292, + "loss": 0.2021, + "step": 33760 + }, + { + "epoch": 1.3039113479284914, + "grad_norm": 1.148047685623169, + "learning_rate": 0.00011307515090672742, + "loss": 0.3987, + "step": 33770 + }, + { + "epoch": 1.3042974632225182, + "grad_norm": 0.19721268117427826, + "learning_rate": 0.00011304940988712562, + "loss": 0.2642, + "step": 33780 + }, + { + "epoch": 1.3046835785165452, + "grad_norm": 1.4060617685317993, + "learning_rate": 0.00011302366886752385, + "loss": 0.2736, + "step": 33790 + }, + { + "epoch": 1.305069693810572, + "grad_norm": 1.0736548900604248, + "learning_rate": 0.00011299792784792207, + "loss": 0.2448, + "step": 33800 + }, + { + "epoch": 1.3054558091045987, + "grad_norm": 4.352476119995117, + "learning_rate": 0.00011297218682832027, + "loss": 0.383, + "step": 33810 + }, + { + "epoch": 1.3058419243986255, + "grad_norm": 0.2249228060245514, + "learning_rate": 0.0001129464458087185, + "loss": 0.14, + "step": 33820 + }, + { + "epoch": 1.3062280396926522, + "grad_norm": 0.4820781946182251, + "learning_rate": 0.0001129207047891167, + "loss": 0.248, + "step": 33830 + }, + { + "epoch": 1.306614154986679, + "grad_norm": 2.2983391284942627, + "learning_rate": 0.00011289496376951491, + "loss": 0.2608, + "step": 33840 + }, + { + "epoch": 1.3070002702807058, + "grad_norm": 1.3315671682357788, + "learning_rate": 0.00011286922274991314, + "loss": 0.1727, + "step": 33850 + }, + { + "epoch": 1.3073863855747327, + "grad_norm": 2.060299873352051, + "learning_rate": 0.00011284348173031134, + "loss": 0.3097, + "step": 33860 + }, + { + "epoch": 1.3077725008687593, + "grad_norm": 2.096285581588745, + "learning_rate": 0.00011281774071070957, + "loss": 0.2305, + "step": 33870 + }, + { + "epoch": 1.3081586161627863, + "grad_norm": 0.4997636675834656, + "learning_rate": 0.00011279199969110777, + "loss": 0.1993, + "step": 33880 + }, + { + "epoch": 1.308544731456813, + "grad_norm": 0.58636474609375, + "learning_rate": 0.00011276625867150599, + "loss": 0.1873, + "step": 33890 + }, + { + "epoch": 1.3089308467508398, + "grad_norm": 0.9128592610359192, + "learning_rate": 0.00011274051765190419, + "loss": 0.1885, + "step": 33900 + }, + { + "epoch": 1.3093169620448666, + "grad_norm": 2.228043794631958, + "learning_rate": 0.0001127147766323024, + "loss": 0.3649, + "step": 33910 + }, + { + "epoch": 1.3097030773388934, + "grad_norm": 1.069002389907837, + "learning_rate": 0.00011268903561270063, + "loss": 0.5454, + "step": 33920 + }, + { + "epoch": 1.3100891926329201, + "grad_norm": 0.6207597851753235, + "learning_rate": 0.00011266329459309883, + "loss": 0.2329, + "step": 33930 + }, + { + "epoch": 1.3104753079269469, + "grad_norm": 1.262247920036316, + "learning_rate": 0.00011263755357349706, + "loss": 0.3437, + "step": 33940 + }, + { + "epoch": 1.3108614232209739, + "grad_norm": 1.7429994344711304, + "learning_rate": 0.00011261181255389526, + "loss": 0.228, + "step": 33950 + }, + { + "epoch": 1.3112475385150006, + "grad_norm": 0.646900475025177, + "learning_rate": 0.00011258607153429349, + "loss": 0.3739, + "step": 33960 + }, + { + "epoch": 1.3116336538090274, + "grad_norm": 1.8228782415390015, + "learning_rate": 0.0001125603305146917, + "loss": 0.2325, + "step": 33970 + }, + { + "epoch": 1.3120197691030542, + "grad_norm": 3.539228916168213, + "learning_rate": 0.0001125345894950899, + "loss": 0.194, + "step": 33980 + }, + { + "epoch": 1.312405884397081, + "grad_norm": 1.2801135778427124, + "learning_rate": 0.00011250884847548813, + "loss": 0.3069, + "step": 33990 + }, + { + "epoch": 1.3127919996911077, + "grad_norm": 3.6265695095062256, + "learning_rate": 0.00011248310745588633, + "loss": 0.3113, + "step": 34000 + }, + { + "epoch": 1.3131781149851345, + "grad_norm": 0.07370063662528992, + "learning_rate": 0.00011245736643628455, + "loss": 0.1449, + "step": 34010 + }, + { + "epoch": 1.3135642302791615, + "grad_norm": 1.0295637845993042, + "learning_rate": 0.00011243162541668275, + "loss": 0.22, + "step": 34020 + }, + { + "epoch": 1.313950345573188, + "grad_norm": 0.8803662061691284, + "learning_rate": 0.00011240588439708098, + "loss": 0.1368, + "step": 34030 + }, + { + "epoch": 1.314336460867215, + "grad_norm": 1.6597707271575928, + "learning_rate": 0.00011238014337747919, + "loss": 0.3038, + "step": 34040 + }, + { + "epoch": 1.3147225761612418, + "grad_norm": 2.115492343902588, + "learning_rate": 0.00011235440235787739, + "loss": 0.1754, + "step": 34050 + }, + { + "epoch": 1.3151086914552685, + "grad_norm": 0.8143919706344604, + "learning_rate": 0.00011232866133827562, + "loss": 0.3764, + "step": 34060 + }, + { + "epoch": 1.3154948067492953, + "grad_norm": 0.14369767904281616, + "learning_rate": 0.00011230292031867382, + "loss": 0.1742, + "step": 34070 + }, + { + "epoch": 1.315880922043322, + "grad_norm": 1.0129845142364502, + "learning_rate": 0.00011227717929907205, + "loss": 0.1458, + "step": 34080 + }, + { + "epoch": 1.316267037337349, + "grad_norm": 2.7300291061401367, + "learning_rate": 0.00011225143827947025, + "loss": 0.3939, + "step": 34090 + }, + { + "epoch": 1.3166531526313756, + "grad_norm": 0.20205609500408173, + "learning_rate": 0.00011222569725986847, + "loss": 0.29, + "step": 34100 + }, + { + "epoch": 1.3170392679254026, + "grad_norm": 1.8928464651107788, + "learning_rate": 0.00011219995624026669, + "loss": 0.1742, + "step": 34110 + }, + { + "epoch": 1.3174253832194294, + "grad_norm": 0.2639687955379486, + "learning_rate": 0.00011217421522066488, + "loss": 0.1745, + "step": 34120 + }, + { + "epoch": 1.3178114985134561, + "grad_norm": 0.5906389355659485, + "learning_rate": 0.00011214847420106311, + "loss": 0.2134, + "step": 34130 + }, + { + "epoch": 1.318197613807483, + "grad_norm": 0.9190629720687866, + "learning_rate": 0.00011212273318146131, + "loss": 0.2547, + "step": 34140 + }, + { + "epoch": 1.3185837291015097, + "grad_norm": 0.5749151110649109, + "learning_rate": 0.00011209699216185954, + "loss": 0.1688, + "step": 34150 + }, + { + "epoch": 1.3189698443955364, + "grad_norm": 0.82295823097229, + "learning_rate": 0.00011207125114225777, + "loss": 0.2884, + "step": 34160 + }, + { + "epoch": 1.3193559596895632, + "grad_norm": 0.07816460728645325, + "learning_rate": 0.00011204551012265597, + "loss": 0.2418, + "step": 34170 + }, + { + "epoch": 1.3197420749835902, + "grad_norm": 0.6417407393455505, + "learning_rate": 0.00011201976910305418, + "loss": 0.2557, + "step": 34180 + }, + { + "epoch": 1.320128190277617, + "grad_norm": 6.093267440795898, + "learning_rate": 0.00011199402808345238, + "loss": 0.3088, + "step": 34190 + }, + { + "epoch": 1.3205143055716437, + "grad_norm": 1.8861887454986572, + "learning_rate": 0.0001119682870638506, + "loss": 0.2204, + "step": 34200 + }, + { + "epoch": 1.3209004208656705, + "grad_norm": 2.3272714614868164, + "learning_rate": 0.0001119425460442488, + "loss": 0.2236, + "step": 34210 + }, + { + "epoch": 1.3212865361596973, + "grad_norm": 0.9608810544013977, + "learning_rate": 0.00011191680502464703, + "loss": 0.1897, + "step": 34220 + }, + { + "epoch": 1.321672651453724, + "grad_norm": 1.2157350778579712, + "learning_rate": 0.00011189106400504526, + "loss": 0.1526, + "step": 34230 + }, + { + "epoch": 1.3220587667477508, + "grad_norm": 1.6684671640396118, + "learning_rate": 0.00011186532298544346, + "loss": 0.3394, + "step": 34240 + }, + { + "epoch": 1.3224448820417778, + "grad_norm": 2.0432374477386475, + "learning_rate": 0.00011183958196584167, + "loss": 0.2183, + "step": 34250 + }, + { + "epoch": 1.3228309973358043, + "grad_norm": 0.9436892867088318, + "learning_rate": 0.00011181384094623988, + "loss": 0.2947, + "step": 34260 + }, + { + "epoch": 1.3232171126298313, + "grad_norm": 0.23260092735290527, + "learning_rate": 0.0001117880999266381, + "loss": 0.114, + "step": 34270 + }, + { + "epoch": 1.323603227923858, + "grad_norm": 1.2291594743728638, + "learning_rate": 0.0001117623589070363, + "loss": 0.3145, + "step": 34280 + }, + { + "epoch": 1.3239893432178849, + "grad_norm": 0.41411107778549194, + "learning_rate": 0.00011173661788743452, + "loss": 0.2937, + "step": 34290 + }, + { + "epoch": 1.3243754585119116, + "grad_norm": 2.354405164718628, + "learning_rate": 0.00011171087686783275, + "loss": 0.3933, + "step": 34300 + }, + { + "epoch": 1.3247615738059384, + "grad_norm": 2.6997978687286377, + "learning_rate": 0.00011168513584823095, + "loss": 0.1494, + "step": 34310 + }, + { + "epoch": 1.3251476890999654, + "grad_norm": 2.8430919647216797, + "learning_rate": 0.00011165939482862916, + "loss": 0.2869, + "step": 34320 + }, + { + "epoch": 1.325533804393992, + "grad_norm": 1.1737356185913086, + "learning_rate": 0.00011163365380902738, + "loss": 0.2792, + "step": 34330 + }, + { + "epoch": 1.325919919688019, + "grad_norm": 4.123973846435547, + "learning_rate": 0.00011160791278942559, + "loss": 0.5211, + "step": 34340 + }, + { + "epoch": 1.3263060349820457, + "grad_norm": 0.8862038850784302, + "learning_rate": 0.00011158217176982382, + "loss": 0.2976, + "step": 34350 + }, + { + "epoch": 1.3266921502760725, + "grad_norm": 1.8690590858459473, + "learning_rate": 0.00011155643075022202, + "loss": 0.2485, + "step": 34360 + }, + { + "epoch": 1.3270782655700992, + "grad_norm": 2.885589599609375, + "learning_rate": 0.00011153068973062024, + "loss": 0.3084, + "step": 34370 + }, + { + "epoch": 1.327464380864126, + "grad_norm": 0.9898788928985596, + "learning_rate": 0.00011150494871101844, + "loss": 0.261, + "step": 34380 + }, + { + "epoch": 1.3278504961581528, + "grad_norm": 0.6879653930664062, + "learning_rate": 0.00011147920769141667, + "loss": 0.2082, + "step": 34390 + }, + { + "epoch": 1.3282366114521795, + "grad_norm": 1.2619003057479858, + "learning_rate": 0.00011145346667181487, + "loss": 0.2402, + "step": 34400 + }, + { + "epoch": 1.3286227267462065, + "grad_norm": 1.1212007999420166, + "learning_rate": 0.00011142772565221308, + "loss": 0.3525, + "step": 34410 + }, + { + "epoch": 1.3290088420402333, + "grad_norm": 1.8431956768035889, + "learning_rate": 0.00011140198463261131, + "loss": 0.2212, + "step": 34420 + }, + { + "epoch": 1.32939495733426, + "grad_norm": 0.6185423731803894, + "learning_rate": 0.00011137624361300951, + "loss": 0.2455, + "step": 34430 + }, + { + "epoch": 1.3297810726282868, + "grad_norm": 2.3791301250457764, + "learning_rate": 0.00011135050259340774, + "loss": 0.1763, + "step": 34440 + }, + { + "epoch": 1.3301671879223136, + "grad_norm": 0.4928603768348694, + "learning_rate": 0.00011132476157380594, + "loss": 0.2381, + "step": 34450 + }, + { + "epoch": 1.3305533032163404, + "grad_norm": 1.5636029243469238, + "learning_rate": 0.00011129902055420416, + "loss": 0.1368, + "step": 34460 + }, + { + "epoch": 1.3309394185103671, + "grad_norm": 0.9425283074378967, + "learning_rate": 0.00011127327953460238, + "loss": 0.2825, + "step": 34470 + }, + { + "epoch": 1.3313255338043941, + "grad_norm": 1.2257115840911865, + "learning_rate": 0.00011124753851500058, + "loss": 0.2547, + "step": 34480 + }, + { + "epoch": 1.3317116490984207, + "grad_norm": 0.9416170716285706, + "learning_rate": 0.0001112217974953988, + "loss": 0.2766, + "step": 34490 + }, + { + "epoch": 1.3320977643924476, + "grad_norm": 0.5123847126960754, + "learning_rate": 0.000111196056475797, + "loss": 0.4733, + "step": 34500 + }, + { + "epoch": 1.3324838796864744, + "grad_norm": 1.5581384897232056, + "learning_rate": 0.00011117031545619523, + "loss": 0.1597, + "step": 34510 + }, + { + "epoch": 1.3328699949805012, + "grad_norm": 2.377333879470825, + "learning_rate": 0.00011114457443659343, + "loss": 0.209, + "step": 34520 + }, + { + "epoch": 1.333256110274528, + "grad_norm": 1.7840913534164429, + "learning_rate": 0.00011111883341699166, + "loss": 0.1759, + "step": 34530 + }, + { + "epoch": 1.3336422255685547, + "grad_norm": 1.1825993061065674, + "learning_rate": 0.00011109309239738987, + "loss": 0.2464, + "step": 34540 + }, + { + "epoch": 1.3340283408625815, + "grad_norm": 1.8859659433364868, + "learning_rate": 0.00011106735137778807, + "loss": 0.3539, + "step": 34550 + }, + { + "epoch": 1.3344144561566083, + "grad_norm": 1.9698175191879272, + "learning_rate": 0.0001110416103581863, + "loss": 0.3301, + "step": 34560 + }, + { + "epoch": 1.3348005714506352, + "grad_norm": 0.7649385333061218, + "learning_rate": 0.0001110158693385845, + "loss": 0.232, + "step": 34570 + }, + { + "epoch": 1.335186686744662, + "grad_norm": 0.56386399269104, + "learning_rate": 0.00011099012831898272, + "loss": 0.3425, + "step": 34580 + }, + { + "epoch": 1.3355728020386888, + "grad_norm": 2.956003189086914, + "learning_rate": 0.00011096438729938092, + "loss": 0.1518, + "step": 34590 + }, + { + "epoch": 1.3359589173327155, + "grad_norm": 2.612029552459717, + "learning_rate": 0.00011093864627977915, + "loss": 0.2765, + "step": 34600 + }, + { + "epoch": 1.3363450326267423, + "grad_norm": 0.9674397706985474, + "learning_rate": 0.00011091290526017736, + "loss": 0.303, + "step": 34610 + }, + { + "epoch": 1.336731147920769, + "grad_norm": 0.9578921794891357, + "learning_rate": 0.00011088716424057556, + "loss": 0.1405, + "step": 34620 + }, + { + "epoch": 1.3371172632147958, + "grad_norm": 2.168065071105957, + "learning_rate": 0.00011086142322097379, + "loss": 0.1914, + "step": 34630 + }, + { + "epoch": 1.3375033785088228, + "grad_norm": 1.3166526556015015, + "learning_rate": 0.00011083568220137199, + "loss": 0.4134, + "step": 34640 + }, + { + "epoch": 1.3378894938028496, + "grad_norm": 0.9082283973693848, + "learning_rate": 0.00011080994118177022, + "loss": 0.2693, + "step": 34650 + }, + { + "epoch": 1.3382756090968764, + "grad_norm": 2.203007698059082, + "learning_rate": 0.00011078420016216844, + "loss": 0.1847, + "step": 34660 + }, + { + "epoch": 1.3386617243909031, + "grad_norm": 0.8101674914360046, + "learning_rate": 0.00011075845914256664, + "loss": 0.3111, + "step": 34670 + }, + { + "epoch": 1.33904783968493, + "grad_norm": 1.9545695781707764, + "learning_rate": 0.00011073271812296486, + "loss": 0.3761, + "step": 34680 + }, + { + "epoch": 1.3394339549789567, + "grad_norm": 1.547581672668457, + "learning_rate": 0.00011070697710336306, + "loss": 0.2374, + "step": 34690 + }, + { + "epoch": 1.3398200702729834, + "grad_norm": 3.3519034385681152, + "learning_rate": 0.00011068123608376128, + "loss": 0.1957, + "step": 34700 + }, + { + "epoch": 1.3402061855670104, + "grad_norm": 1.5508599281311035, + "learning_rate": 0.00011065549506415948, + "loss": 0.4171, + "step": 34710 + }, + { + "epoch": 1.340592300861037, + "grad_norm": 1.8547546863555908, + "learning_rate": 0.00011062975404455771, + "loss": 0.1872, + "step": 34720 + }, + { + "epoch": 1.340978416155064, + "grad_norm": 1.4600756168365479, + "learning_rate": 0.00011060401302495594, + "loss": 0.3515, + "step": 34730 + }, + { + "epoch": 1.3413645314490907, + "grad_norm": 0.05774044618010521, + "learning_rate": 0.00011057827200535414, + "loss": 0.1604, + "step": 34740 + }, + { + "epoch": 1.3417506467431175, + "grad_norm": 2.8793342113494873, + "learning_rate": 0.00011055253098575235, + "loss": 0.3095, + "step": 34750 + }, + { + "epoch": 1.3421367620371443, + "grad_norm": 2.241042375564575, + "learning_rate": 0.00011052678996615055, + "loss": 0.2511, + "step": 34760 + }, + { + "epoch": 1.342522877331171, + "grad_norm": 1.9320632219314575, + "learning_rate": 0.00011050104894654878, + "loss": 0.4493, + "step": 34770 + }, + { + "epoch": 1.3429089926251978, + "grad_norm": 1.6483882665634155, + "learning_rate": 0.000110475307926947, + "loss": 0.217, + "step": 34780 + }, + { + "epoch": 1.3432951079192246, + "grad_norm": 0.9635765552520752, + "learning_rate": 0.0001104495669073452, + "loss": 0.5458, + "step": 34790 + }, + { + "epoch": 1.3436812232132516, + "grad_norm": 1.2436567544937134, + "learning_rate": 0.00011042382588774343, + "loss": 0.2857, + "step": 34800 + }, + { + "epoch": 1.3440673385072783, + "grad_norm": 2.8082425594329834, + "learning_rate": 0.00011039808486814163, + "loss": 0.3439, + "step": 34810 + }, + { + "epoch": 1.344453453801305, + "grad_norm": 1.0430901050567627, + "learning_rate": 0.00011037234384853984, + "loss": 0.1404, + "step": 34820 + }, + { + "epoch": 1.3448395690953319, + "grad_norm": 1.7387149333953857, + "learning_rate": 0.00011034660282893806, + "loss": 0.395, + "step": 34830 + }, + { + "epoch": 1.3452256843893586, + "grad_norm": 1.2713748216629028, + "learning_rate": 0.00011032086180933627, + "loss": 0.29, + "step": 34840 + }, + { + "epoch": 1.3456117996833854, + "grad_norm": 0.26068204641342163, + "learning_rate": 0.0001102951207897345, + "loss": 0.1814, + "step": 34850 + }, + { + "epoch": 1.3459979149774122, + "grad_norm": 2.4163243770599365, + "learning_rate": 0.0001102693797701327, + "loss": 0.2445, + "step": 34860 + }, + { + "epoch": 1.3463840302714392, + "grad_norm": 2.2439687252044678, + "learning_rate": 0.00011024363875053092, + "loss": 0.3338, + "step": 34870 + }, + { + "epoch": 1.346770145565466, + "grad_norm": 0.2822403609752655, + "learning_rate": 0.00011021789773092912, + "loss": 0.1648, + "step": 34880 + }, + { + "epoch": 1.3471562608594927, + "grad_norm": 0.07319017499685287, + "learning_rate": 0.00011019215671132734, + "loss": 0.107, + "step": 34890 + }, + { + "epoch": 1.3475423761535195, + "grad_norm": 0.9809044003486633, + "learning_rate": 0.00011016641569172555, + "loss": 0.256, + "step": 34900 + }, + { + "epoch": 1.3479284914475462, + "grad_norm": 0.5016226768493652, + "learning_rate": 0.00011014067467212376, + "loss": 0.3229, + "step": 34910 + }, + { + "epoch": 1.348314606741573, + "grad_norm": 1.3026005029678345, + "learning_rate": 0.00011011493365252199, + "loss": 0.2145, + "step": 34920 + }, + { + "epoch": 1.3487007220355998, + "grad_norm": 1.0752215385437012, + "learning_rate": 0.00011008919263292019, + "loss": 0.2355, + "step": 34930 + }, + { + "epoch": 1.3490868373296268, + "grad_norm": 2.2703003883361816, + "learning_rate": 0.00011006345161331842, + "loss": 0.2079, + "step": 34940 + }, + { + "epoch": 1.3494729526236533, + "grad_norm": 1.1323810815811157, + "learning_rate": 0.00011003771059371662, + "loss": 0.1015, + "step": 34950 + }, + { + "epoch": 1.3498590679176803, + "grad_norm": 0.10813555121421814, + "learning_rate": 0.00011001196957411484, + "loss": 0.4214, + "step": 34960 + }, + { + "epoch": 1.350245183211707, + "grad_norm": 0.07815568149089813, + "learning_rate": 0.00010998622855451306, + "loss": 0.1211, + "step": 34970 + }, + { + "epoch": 1.3506312985057338, + "grad_norm": 0.6748234629631042, + "learning_rate": 0.00010996048753491126, + "loss": 0.3508, + "step": 34980 + }, + { + "epoch": 1.3510174137997606, + "grad_norm": 1.8556997776031494, + "learning_rate": 0.00010993474651530948, + "loss": 0.2268, + "step": 34990 + }, + { + "epoch": 1.3514035290937874, + "grad_norm": 0.8696061372756958, + "learning_rate": 0.00010990900549570768, + "loss": 0.4321, + "step": 35000 + }, + { + "epoch": 1.3517896443878141, + "grad_norm": 0.42442765831947327, + "learning_rate": 0.00010988326447610591, + "loss": 0.1944, + "step": 35010 + }, + { + "epoch": 1.352175759681841, + "grad_norm": 1.0474554300308228, + "learning_rate": 0.00010985752345650411, + "loss": 0.1342, + "step": 35020 + }, + { + "epoch": 1.3525618749758679, + "grad_norm": 0.607037365436554, + "learning_rate": 0.00010983178243690234, + "loss": 0.2965, + "step": 35030 + }, + { + "epoch": 1.3529479902698947, + "grad_norm": 1.8160990476608276, + "learning_rate": 0.00010980604141730055, + "loss": 0.3192, + "step": 35040 + }, + { + "epoch": 1.3533341055639214, + "grad_norm": 2.0026509761810303, + "learning_rate": 0.00010978030039769875, + "loss": 0.3054, + "step": 35050 + }, + { + "epoch": 1.3537202208579482, + "grad_norm": 0.9203600883483887, + "learning_rate": 0.00010975455937809698, + "loss": 0.253, + "step": 35060 + }, + { + "epoch": 1.354106336151975, + "grad_norm": 0.33198195695877075, + "learning_rate": 0.00010972881835849518, + "loss": 0.3885, + "step": 35070 + }, + { + "epoch": 1.3544924514460017, + "grad_norm": 0.3201223611831665, + "learning_rate": 0.0001097030773388934, + "loss": 0.3029, + "step": 35080 + }, + { + "epoch": 1.3548785667400285, + "grad_norm": 1.2589943408966064, + "learning_rate": 0.0001096773363192916, + "loss": 0.4243, + "step": 35090 + }, + { + "epoch": 1.3552646820340555, + "grad_norm": 1.5106219053268433, + "learning_rate": 0.00010965159529968983, + "loss": 0.2585, + "step": 35100 + }, + { + "epoch": 1.3556507973280822, + "grad_norm": 1.429799199104309, + "learning_rate": 0.00010962585428008804, + "loss": 0.1961, + "step": 35110 + }, + { + "epoch": 1.356036912622109, + "grad_norm": 2.1211297512054443, + "learning_rate": 0.00010960011326048624, + "loss": 0.4057, + "step": 35120 + }, + { + "epoch": 1.3564230279161358, + "grad_norm": 2.5154731273651123, + "learning_rate": 0.00010957437224088447, + "loss": 0.3787, + "step": 35130 + }, + { + "epoch": 1.3568091432101625, + "grad_norm": 0.4914834201335907, + "learning_rate": 0.00010954863122128267, + "loss": 0.234, + "step": 35140 + }, + { + "epoch": 1.3571952585041893, + "grad_norm": 0.26685893535614014, + "learning_rate": 0.0001095228902016809, + "loss": 0.2841, + "step": 35150 + }, + { + "epoch": 1.357581373798216, + "grad_norm": 0.15462155640125275, + "learning_rate": 0.00010949714918207912, + "loss": 0.2269, + "step": 35160 + }, + { + "epoch": 1.357967489092243, + "grad_norm": 1.3887063264846802, + "learning_rate": 0.00010947140816247732, + "loss": 0.3455, + "step": 35170 + }, + { + "epoch": 1.3583536043862696, + "grad_norm": 0.786374032497406, + "learning_rate": 0.00010944566714287554, + "loss": 0.2897, + "step": 35180 + }, + { + "epoch": 1.3587397196802966, + "grad_norm": 1.100475549697876, + "learning_rate": 0.00010941992612327374, + "loss": 0.2892, + "step": 35190 + }, + { + "epoch": 1.3591258349743234, + "grad_norm": 0.7676102519035339, + "learning_rate": 0.00010939418510367196, + "loss": 0.1942, + "step": 35200 + }, + { + "epoch": 1.3595119502683501, + "grad_norm": 0.33462053537368774, + "learning_rate": 0.00010936844408407016, + "loss": 0.2872, + "step": 35210 + }, + { + "epoch": 1.359898065562377, + "grad_norm": 0.9294387698173523, + "learning_rate": 0.00010934270306446839, + "loss": 0.2617, + "step": 35220 + }, + { + "epoch": 1.3602841808564037, + "grad_norm": 0.3169979453086853, + "learning_rate": 0.00010931696204486662, + "loss": 0.2942, + "step": 35230 + }, + { + "epoch": 1.3606702961504304, + "grad_norm": 2.1339616775512695, + "learning_rate": 0.00010929122102526482, + "loss": 0.4448, + "step": 35240 + }, + { + "epoch": 1.3610564114444572, + "grad_norm": 0.9430062770843506, + "learning_rate": 0.00010926548000566303, + "loss": 0.2051, + "step": 35250 + }, + { + "epoch": 1.3614425267384842, + "grad_norm": 3.1187360286712646, + "learning_rate": 0.00010923973898606123, + "loss": 0.2274, + "step": 35260 + }, + { + "epoch": 1.361828642032511, + "grad_norm": 1.4727579355239868, + "learning_rate": 0.00010921399796645946, + "loss": 0.3757, + "step": 35270 + }, + { + "epoch": 1.3622147573265377, + "grad_norm": 2.157560348510742, + "learning_rate": 0.00010918825694685768, + "loss": 0.3096, + "step": 35280 + }, + { + "epoch": 1.3626008726205645, + "grad_norm": 0.33457377552986145, + "learning_rate": 0.00010916251592725588, + "loss": 0.1489, + "step": 35290 + }, + { + "epoch": 1.3629869879145913, + "grad_norm": 0.9005904197692871, + "learning_rate": 0.00010913677490765411, + "loss": 0.1826, + "step": 35300 + }, + { + "epoch": 1.363373103208618, + "grad_norm": 2.1222829818725586, + "learning_rate": 0.00010911103388805231, + "loss": 0.1965, + "step": 35310 + }, + { + "epoch": 1.3637592185026448, + "grad_norm": 1.3881357908248901, + "learning_rate": 0.00010908529286845052, + "loss": 0.1791, + "step": 35320 + }, + { + "epoch": 1.3641453337966718, + "grad_norm": 1.7574503421783447, + "learning_rate": 0.00010905955184884872, + "loss": 0.3316, + "step": 35330 + }, + { + "epoch": 1.3645314490906983, + "grad_norm": 0.1967727392911911, + "learning_rate": 0.00010903381082924695, + "loss": 0.2331, + "step": 35340 + }, + { + "epoch": 1.3649175643847253, + "grad_norm": 0.8974360823631287, + "learning_rate": 0.00010900806980964518, + "loss": 0.2589, + "step": 35350 + }, + { + "epoch": 1.365303679678752, + "grad_norm": 2.0996744632720947, + "learning_rate": 0.00010898232879004338, + "loss": 0.3663, + "step": 35360 + }, + { + "epoch": 1.3656897949727789, + "grad_norm": 0.5678316354751587, + "learning_rate": 0.0001089565877704416, + "loss": 0.1729, + "step": 35370 + }, + { + "epoch": 1.3660759102668056, + "grad_norm": 2.3381874561309814, + "learning_rate": 0.0001089308467508398, + "loss": 0.1615, + "step": 35380 + }, + { + "epoch": 1.3664620255608324, + "grad_norm": 1.0276836156845093, + "learning_rate": 0.00010890510573123802, + "loss": 0.3359, + "step": 35390 + }, + { + "epoch": 1.3668481408548594, + "grad_norm": 2.4374940395355225, + "learning_rate": 0.00010887936471163622, + "loss": 0.2435, + "step": 35400 + }, + { + "epoch": 1.367234256148886, + "grad_norm": 0.45221665501594543, + "learning_rate": 0.00010885362369203444, + "loss": 0.2555, + "step": 35410 + }, + { + "epoch": 1.367620371442913, + "grad_norm": 2.608090400695801, + "learning_rate": 0.00010882788267243267, + "loss": 0.2465, + "step": 35420 + }, + { + "epoch": 1.3680064867369397, + "grad_norm": 1.4186642169952393, + "learning_rate": 0.00010880214165283087, + "loss": 0.1674, + "step": 35430 + }, + { + "epoch": 1.3683926020309665, + "grad_norm": 0.659479022026062, + "learning_rate": 0.0001087764006332291, + "loss": 0.2926, + "step": 35440 + }, + { + "epoch": 1.3687787173249932, + "grad_norm": 0.9219567179679871, + "learning_rate": 0.0001087506596136273, + "loss": 0.2001, + "step": 35450 + }, + { + "epoch": 1.36916483261902, + "grad_norm": 0.8070804476737976, + "learning_rate": 0.00010872491859402551, + "loss": 0.2178, + "step": 35460 + }, + { + "epoch": 1.3695509479130468, + "grad_norm": 2.9981069564819336, + "learning_rate": 0.00010869917757442374, + "loss": 0.3079, + "step": 35470 + }, + { + "epoch": 1.3699370632070735, + "grad_norm": 0.7891242504119873, + "learning_rate": 0.00010867343655482194, + "loss": 0.2765, + "step": 35480 + }, + { + "epoch": 1.3703231785011005, + "grad_norm": 1.448637843132019, + "learning_rate": 0.00010864769553522016, + "loss": 0.3521, + "step": 35490 + }, + { + "epoch": 1.3707092937951273, + "grad_norm": 0.07628043740987778, + "learning_rate": 0.00010862195451561836, + "loss": 0.2083, + "step": 35500 + }, + { + "epoch": 1.371095409089154, + "grad_norm": 0.7549735307693481, + "learning_rate": 0.00010859621349601659, + "loss": 0.2536, + "step": 35510 + }, + { + "epoch": 1.3714815243831808, + "grad_norm": 1.3548041582107544, + "learning_rate": 0.00010857047247641479, + "loss": 0.251, + "step": 35520 + }, + { + "epoch": 1.3718676396772076, + "grad_norm": 0.530010998249054, + "learning_rate": 0.000108544731456813, + "loss": 0.1917, + "step": 35530 + }, + { + "epoch": 1.3722537549712344, + "grad_norm": 0.4148992896080017, + "learning_rate": 0.00010851899043721123, + "loss": 0.335, + "step": 35540 + }, + { + "epoch": 1.3726398702652611, + "grad_norm": 1.5118776559829712, + "learning_rate": 0.00010849324941760943, + "loss": 0.2159, + "step": 35550 + }, + { + "epoch": 1.3730259855592881, + "grad_norm": 1.036889910697937, + "learning_rate": 0.00010846750839800766, + "loss": 0.2975, + "step": 35560 + }, + { + "epoch": 1.3734121008533147, + "grad_norm": 1.724263072013855, + "learning_rate": 0.00010844176737840586, + "loss": 0.1476, + "step": 35570 + }, + { + "epoch": 1.3737982161473417, + "grad_norm": 1.599007487297058, + "learning_rate": 0.00010841602635880408, + "loss": 0.2539, + "step": 35580 + }, + { + "epoch": 1.3741843314413684, + "grad_norm": 2.9119279384613037, + "learning_rate": 0.00010839028533920228, + "loss": 0.2688, + "step": 35590 + }, + { + "epoch": 1.3745704467353952, + "grad_norm": 1.8647874593734741, + "learning_rate": 0.00010836454431960051, + "loss": 0.4158, + "step": 35600 + }, + { + "epoch": 1.374956562029422, + "grad_norm": 3.925290822982788, + "learning_rate": 0.00010833880329999872, + "loss": 0.3333, + "step": 35610 + }, + { + "epoch": 1.3753426773234487, + "grad_norm": 0.7124634385108948, + "learning_rate": 0.00010831306228039692, + "loss": 0.1069, + "step": 35620 + }, + { + "epoch": 1.3757287926174757, + "grad_norm": 1.303579330444336, + "learning_rate": 0.00010828732126079515, + "loss": 0.2898, + "step": 35630 + }, + { + "epoch": 1.3761149079115023, + "grad_norm": 3.921804189682007, + "learning_rate": 0.00010826158024119335, + "loss": 0.4212, + "step": 35640 + }, + { + "epoch": 1.3765010232055293, + "grad_norm": 1.3194564580917358, + "learning_rate": 0.00010823583922159158, + "loss": 0.2771, + "step": 35650 + }, + { + "epoch": 1.376887138499556, + "grad_norm": 1.4237637519836426, + "learning_rate": 0.00010821009820198979, + "loss": 0.2463, + "step": 35660 + }, + { + "epoch": 1.3772732537935828, + "grad_norm": 1.8165888786315918, + "learning_rate": 0.000108184357182388, + "loss": 0.291, + "step": 35670 + }, + { + "epoch": 1.3776593690876096, + "grad_norm": 1.1056426763534546, + "learning_rate": 0.00010815861616278622, + "loss": 0.2525, + "step": 35680 + }, + { + "epoch": 1.3780454843816363, + "grad_norm": 1.483189582824707, + "learning_rate": 0.00010813287514318442, + "loss": 0.1569, + "step": 35690 + }, + { + "epoch": 1.378431599675663, + "grad_norm": 1.0666841268539429, + "learning_rate": 0.00010810713412358264, + "loss": 0.235, + "step": 35700 + }, + { + "epoch": 1.3788177149696899, + "grad_norm": 1.0299845933914185, + "learning_rate": 0.00010808139310398084, + "loss": 0.3892, + "step": 35710 + }, + { + "epoch": 1.3792038302637168, + "grad_norm": 2.3474409580230713, + "learning_rate": 0.00010805565208437907, + "loss": 0.3417, + "step": 35720 + }, + { + "epoch": 1.3795899455577436, + "grad_norm": 1.7456315755844116, + "learning_rate": 0.0001080299110647773, + "loss": 0.2538, + "step": 35730 + }, + { + "epoch": 1.3799760608517704, + "grad_norm": 2.866103410720825, + "learning_rate": 0.0001080041700451755, + "loss": 0.1619, + "step": 35740 + }, + { + "epoch": 1.3803621761457971, + "grad_norm": 0.29136407375335693, + "learning_rate": 0.00010797842902557371, + "loss": 0.2692, + "step": 35750 + }, + { + "epoch": 1.380748291439824, + "grad_norm": 0.8046161532402039, + "learning_rate": 0.00010795268800597191, + "loss": 0.1575, + "step": 35760 + }, + { + "epoch": 1.3811344067338507, + "grad_norm": 0.6451787352561951, + "learning_rate": 0.00010792694698637014, + "loss": 0.4914, + "step": 35770 + }, + { + "epoch": 1.3815205220278775, + "grad_norm": 0.7289161086082458, + "learning_rate": 0.00010790120596676836, + "loss": 0.1895, + "step": 35780 + }, + { + "epoch": 1.3819066373219044, + "grad_norm": 0.8300430178642273, + "learning_rate": 0.00010787546494716656, + "loss": 0.3663, + "step": 35790 + }, + { + "epoch": 1.382292752615931, + "grad_norm": 0.17713364958763123, + "learning_rate": 0.00010784972392756479, + "loss": 0.3189, + "step": 35800 + }, + { + "epoch": 1.382678867909958, + "grad_norm": 0.903222918510437, + "learning_rate": 0.00010782398290796299, + "loss": 0.1577, + "step": 35810 + }, + { + "epoch": 1.3830649832039847, + "grad_norm": 0.08617932349443436, + "learning_rate": 0.0001077982418883612, + "loss": 0.2872, + "step": 35820 + }, + { + "epoch": 1.3834510984980115, + "grad_norm": 1.9590895175933838, + "learning_rate": 0.0001077725008687594, + "loss": 0.2907, + "step": 35830 + }, + { + "epoch": 1.3838372137920383, + "grad_norm": 1.2515161037445068, + "learning_rate": 0.00010774675984915763, + "loss": 0.177, + "step": 35840 + }, + { + "epoch": 1.384223329086065, + "grad_norm": 1.6171292066574097, + "learning_rate": 0.00010772101882955586, + "loss": 0.2321, + "step": 35850 + }, + { + "epoch": 1.3846094443800918, + "grad_norm": 0.13681405782699585, + "learning_rate": 0.00010769527780995406, + "loss": 0.247, + "step": 35860 + }, + { + "epoch": 1.3849955596741186, + "grad_norm": 1.1949968338012695, + "learning_rate": 0.00010766953679035228, + "loss": 0.2978, + "step": 35870 + }, + { + "epoch": 1.3853816749681456, + "grad_norm": 0.17001692950725555, + "learning_rate": 0.00010764379577075048, + "loss": 0.3241, + "step": 35880 + }, + { + "epoch": 1.3857677902621723, + "grad_norm": 0.8227952122688293, + "learning_rate": 0.0001076180547511487, + "loss": 0.3499, + "step": 35890 + }, + { + "epoch": 1.386153905556199, + "grad_norm": 1.4185482263565063, + "learning_rate": 0.0001075923137315469, + "loss": 0.3109, + "step": 35900 + }, + { + "epoch": 1.3865400208502259, + "grad_norm": 0.9533351063728333, + "learning_rate": 0.00010756657271194512, + "loss": 0.24, + "step": 35910 + }, + { + "epoch": 1.3869261361442526, + "grad_norm": 1.077789306640625, + "learning_rate": 0.00010754083169234335, + "loss": 0.2662, + "step": 35920 + }, + { + "epoch": 1.3873122514382794, + "grad_norm": 1.3528363704681396, + "learning_rate": 0.00010751509067274155, + "loss": 0.1623, + "step": 35930 + }, + { + "epoch": 1.3876983667323062, + "grad_norm": 0.25122806429862976, + "learning_rate": 0.00010748934965313978, + "loss": 0.1639, + "step": 35940 + }, + { + "epoch": 1.3880844820263332, + "grad_norm": 0.9446159601211548, + "learning_rate": 0.00010746360863353798, + "loss": 0.2035, + "step": 35950 + }, + { + "epoch": 1.38847059732036, + "grad_norm": 1.2258719205856323, + "learning_rate": 0.00010743786761393619, + "loss": 0.4247, + "step": 35960 + }, + { + "epoch": 1.3888567126143867, + "grad_norm": 3.117729663848877, + "learning_rate": 0.00010741212659433442, + "loss": 0.2569, + "step": 35970 + }, + { + "epoch": 1.3892428279084135, + "grad_norm": 0.6333123445510864, + "learning_rate": 0.00010738638557473262, + "loss": 0.1843, + "step": 35980 + }, + { + "epoch": 1.3896289432024402, + "grad_norm": 1.49360191822052, + "learning_rate": 0.00010736064455513084, + "loss": 0.2318, + "step": 35990 + }, + { + "epoch": 1.390015058496467, + "grad_norm": 3.9082753658294678, + "learning_rate": 0.00010733490353552904, + "loss": 0.2037, + "step": 36000 + }, + { + "epoch": 1.3904011737904938, + "grad_norm": 0.5687323808670044, + "learning_rate": 0.00010730916251592727, + "loss": 0.0871, + "step": 36010 + }, + { + "epoch": 1.3907872890845208, + "grad_norm": 0.3480868339538574, + "learning_rate": 0.00010728342149632547, + "loss": 0.4009, + "step": 36020 + }, + { + "epoch": 1.3911734043785473, + "grad_norm": 1.325042486190796, + "learning_rate": 0.00010725768047672368, + "loss": 0.5143, + "step": 36030 + }, + { + "epoch": 1.3915595196725743, + "grad_norm": 2.114786386489868, + "learning_rate": 0.00010723193945712191, + "loss": 0.2944, + "step": 36040 + }, + { + "epoch": 1.391945634966601, + "grad_norm": 1.716272234916687, + "learning_rate": 0.00010720619843752011, + "loss": 0.3793, + "step": 36050 + }, + { + "epoch": 1.3923317502606278, + "grad_norm": 2.057535171508789, + "learning_rate": 0.00010718045741791834, + "loss": 0.199, + "step": 36060 + }, + { + "epoch": 1.3927178655546546, + "grad_norm": 2.371248483657837, + "learning_rate": 0.00010715471639831653, + "loss": 0.3276, + "step": 36070 + }, + { + "epoch": 1.3931039808486814, + "grad_norm": 0.1784186065196991, + "learning_rate": 0.00010712897537871476, + "loss": 0.0922, + "step": 36080 + }, + { + "epoch": 1.3934900961427081, + "grad_norm": 0.7923040390014648, + "learning_rate": 0.00010710323435911298, + "loss": 0.1599, + "step": 36090 + }, + { + "epoch": 1.393876211436735, + "grad_norm": 1.9013831615447998, + "learning_rate": 0.00010707749333951117, + "loss": 0.4997, + "step": 36100 + }, + { + "epoch": 1.394262326730762, + "grad_norm": 1.8659415245056152, + "learning_rate": 0.0001070517523199094, + "loss": 0.1851, + "step": 36110 + }, + { + "epoch": 1.3946484420247887, + "grad_norm": 2.5775375366210938, + "learning_rate": 0.0001070260113003076, + "loss": 0.2384, + "step": 36120 + }, + { + "epoch": 1.3950345573188154, + "grad_norm": 0.21943879127502441, + "learning_rate": 0.00010700027028070583, + "loss": 0.3996, + "step": 36130 + }, + { + "epoch": 1.3954206726128422, + "grad_norm": 1.1734743118286133, + "learning_rate": 0.00010697452926110403, + "loss": 0.1737, + "step": 36140 + }, + { + "epoch": 1.395806787906869, + "grad_norm": 0.697695791721344, + "learning_rate": 0.00010694878824150225, + "loss": 0.2778, + "step": 36150 + }, + { + "epoch": 1.3961929032008957, + "grad_norm": 3.2881579399108887, + "learning_rate": 0.00010692304722190047, + "loss": 0.3198, + "step": 36160 + }, + { + "epoch": 1.3965790184949225, + "grad_norm": 0.1592467725276947, + "learning_rate": 0.00010689730620229868, + "loss": 0.2591, + "step": 36170 + }, + { + "epoch": 1.3969651337889495, + "grad_norm": 0.4579029083251953, + "learning_rate": 0.0001068715651826969, + "loss": 0.2984, + "step": 36180 + }, + { + "epoch": 1.3973512490829763, + "grad_norm": 0.4286015033721924, + "learning_rate": 0.0001068458241630951, + "loss": 0.4098, + "step": 36190 + }, + { + "epoch": 1.397737364377003, + "grad_norm": 1.7824127674102783, + "learning_rate": 0.00010682008314349332, + "loss": 0.2446, + "step": 36200 + }, + { + "epoch": 1.3981234796710298, + "grad_norm": 0.8584449887275696, + "learning_rate": 0.00010679434212389152, + "loss": 0.1415, + "step": 36210 + }, + { + "epoch": 1.3985095949650566, + "grad_norm": 1.1699339151382446, + "learning_rate": 0.00010676860110428975, + "loss": 0.1706, + "step": 36220 + }, + { + "epoch": 1.3988957102590833, + "grad_norm": 2.615877389907837, + "learning_rate": 0.00010674286008468796, + "loss": 0.2605, + "step": 36230 + }, + { + "epoch": 1.39928182555311, + "grad_norm": 2.182037591934204, + "learning_rate": 0.00010671711906508617, + "loss": 0.2067, + "step": 36240 + }, + { + "epoch": 1.399667940847137, + "grad_norm": 2.183263063430786, + "learning_rate": 0.00010669137804548439, + "loss": 0.4704, + "step": 36250 + }, + { + "epoch": 1.4000540561411636, + "grad_norm": 3.505791187286377, + "learning_rate": 0.00010666563702588259, + "loss": 0.3385, + "step": 36260 + }, + { + "epoch": 1.4004401714351906, + "grad_norm": 1.2262030839920044, + "learning_rate": 0.00010663989600628081, + "loss": 0.3999, + "step": 36270 + }, + { + "epoch": 1.4008262867292174, + "grad_norm": 2.4024577140808105, + "learning_rate": 0.00010661415498667904, + "loss": 0.1984, + "step": 36280 + }, + { + "epoch": 1.4012124020232442, + "grad_norm": 0.4166090786457062, + "learning_rate": 0.00010658841396707724, + "loss": 0.2448, + "step": 36290 + }, + { + "epoch": 1.401598517317271, + "grad_norm": 0.422590047121048, + "learning_rate": 0.00010656267294747547, + "loss": 0.2464, + "step": 36300 + }, + { + "epoch": 1.4019846326112977, + "grad_norm": 2.287503480911255, + "learning_rate": 0.00010653693192787367, + "loss": 0.1621, + "step": 36310 + }, + { + "epoch": 1.4023707479053245, + "grad_norm": 0.8126110434532166, + "learning_rate": 0.00010651119090827188, + "loss": 0.2097, + "step": 36320 + }, + { + "epoch": 1.4027568631993512, + "grad_norm": 0.683016836643219, + "learning_rate": 0.00010648544988867008, + "loss": 0.1512, + "step": 36330 + }, + { + "epoch": 1.4031429784933782, + "grad_norm": 1.5477893352508545, + "learning_rate": 0.00010645970886906831, + "loss": 0.2169, + "step": 36340 + }, + { + "epoch": 1.403529093787405, + "grad_norm": 2.183166265487671, + "learning_rate": 0.00010643396784946653, + "loss": 0.4307, + "step": 36350 + }, + { + "epoch": 1.4039152090814317, + "grad_norm": 1.5782747268676758, + "learning_rate": 0.00010640822682986473, + "loss": 0.291, + "step": 36360 + }, + { + "epoch": 1.4043013243754585, + "grad_norm": 1.2823392152786255, + "learning_rate": 0.00010638248581026296, + "loss": 0.315, + "step": 36370 + }, + { + "epoch": 1.4046874396694853, + "grad_norm": 2.0126500129699707, + "learning_rate": 0.00010635674479066116, + "loss": 0.2565, + "step": 36380 + }, + { + "epoch": 1.405073554963512, + "grad_norm": 2.490217447280884, + "learning_rate": 0.00010633100377105937, + "loss": 0.2665, + "step": 36390 + }, + { + "epoch": 1.4054596702575388, + "grad_norm": 1.2054855823516846, + "learning_rate": 0.00010630526275145757, + "loss": 0.5455, + "step": 36400 + }, + { + "epoch": 1.4058457855515658, + "grad_norm": 0.7968757748603821, + "learning_rate": 0.0001062795217318558, + "loss": 0.333, + "step": 36410 + }, + { + "epoch": 1.4062319008455926, + "grad_norm": 1.1027718782424927, + "learning_rate": 0.00010625378071225403, + "loss": 0.2019, + "step": 36420 + }, + { + "epoch": 1.4066180161396193, + "grad_norm": 1.987302541732788, + "learning_rate": 0.00010622803969265223, + "loss": 0.3159, + "step": 36430 + }, + { + "epoch": 1.407004131433646, + "grad_norm": 1.5426512956619263, + "learning_rate": 0.00010620229867305045, + "loss": 0.3759, + "step": 36440 + }, + { + "epoch": 1.4073902467276729, + "grad_norm": 0.5228156447410583, + "learning_rate": 0.00010617655765344865, + "loss": 0.2465, + "step": 36450 + }, + { + "epoch": 1.4077763620216996, + "grad_norm": 0.46890121698379517, + "learning_rate": 0.00010615081663384687, + "loss": 0.302, + "step": 36460 + }, + { + "epoch": 1.4081624773157264, + "grad_norm": 2.1506495475769043, + "learning_rate": 0.0001061250756142451, + "loss": 0.2569, + "step": 36470 + }, + { + "epoch": 1.4085485926097534, + "grad_norm": 2.307468891143799, + "learning_rate": 0.0001060993345946433, + "loss": 0.2009, + "step": 36480 + }, + { + "epoch": 1.40893470790378, + "grad_norm": 0.07033026963472366, + "learning_rate": 0.00010607359357504152, + "loss": 0.1728, + "step": 36490 + }, + { + "epoch": 1.409320823197807, + "grad_norm": 0.3262972831726074, + "learning_rate": 0.00010604785255543972, + "loss": 0.2905, + "step": 36500 + }, + { + "epoch": 1.4097069384918337, + "grad_norm": 0.755646824836731, + "learning_rate": 0.00010602211153583795, + "loss": 0.3287, + "step": 36510 + }, + { + "epoch": 1.4100930537858605, + "grad_norm": 1.1180161237716675, + "learning_rate": 0.00010599637051623615, + "loss": 0.2743, + "step": 36520 + }, + { + "epoch": 1.4104791690798872, + "grad_norm": 1.2358392477035522, + "learning_rate": 0.00010597062949663436, + "loss": 0.2672, + "step": 36530 + }, + { + "epoch": 1.410865284373914, + "grad_norm": 1.4188988208770752, + "learning_rate": 0.00010594488847703259, + "loss": 0.2552, + "step": 36540 + }, + { + "epoch": 1.4112513996679408, + "grad_norm": 1.4727978706359863, + "learning_rate": 0.00010591914745743079, + "loss": 0.2256, + "step": 36550 + }, + { + "epoch": 1.4116375149619675, + "grad_norm": 0.08973213285207748, + "learning_rate": 0.00010589340643782901, + "loss": 0.224, + "step": 36560 + }, + { + "epoch": 1.4120236302559945, + "grad_norm": 0.9915102124214172, + "learning_rate": 0.00010586766541822721, + "loss": 0.195, + "step": 36570 + }, + { + "epoch": 1.4124097455500213, + "grad_norm": 0.8524800539016724, + "learning_rate": 0.00010584192439862544, + "loss": 0.1492, + "step": 36580 + }, + { + "epoch": 1.412795860844048, + "grad_norm": 1.8414466381072998, + "learning_rate": 0.00010581618337902365, + "loss": 0.3412, + "step": 36590 + }, + { + "epoch": 1.4131819761380748, + "grad_norm": 2.596547842025757, + "learning_rate": 0.00010579044235942185, + "loss": 0.3469, + "step": 36600 + }, + { + "epoch": 1.4135680914321016, + "grad_norm": 0.5367813110351562, + "learning_rate": 0.00010576470133982008, + "loss": 0.2274, + "step": 36610 + }, + { + "epoch": 1.4139542067261284, + "grad_norm": 2.70858097076416, + "learning_rate": 0.00010573896032021828, + "loss": 0.171, + "step": 36620 + }, + { + "epoch": 1.4143403220201551, + "grad_norm": 1.7077667713165283, + "learning_rate": 0.00010571321930061651, + "loss": 0.2638, + "step": 36630 + }, + { + "epoch": 1.4147264373141821, + "grad_norm": 0.9189953804016113, + "learning_rate": 0.00010568747828101471, + "loss": 0.2283, + "step": 36640 + }, + { + "epoch": 1.4151125526082087, + "grad_norm": 2.0240087509155273, + "learning_rate": 0.00010566173726141293, + "loss": 0.3166, + "step": 36650 + }, + { + "epoch": 1.4154986679022357, + "grad_norm": 3.3304507732391357, + "learning_rate": 0.00010563599624181115, + "loss": 0.3046, + "step": 36660 + }, + { + "epoch": 1.4158847831962624, + "grad_norm": 1.955929160118103, + "learning_rate": 0.00010561025522220935, + "loss": 0.2653, + "step": 36670 + }, + { + "epoch": 1.4162708984902892, + "grad_norm": 2.2465381622314453, + "learning_rate": 0.00010558451420260757, + "loss": 0.3119, + "step": 36680 + }, + { + "epoch": 1.416657013784316, + "grad_norm": 2.108614921569824, + "learning_rate": 0.00010555877318300577, + "loss": 0.2872, + "step": 36690 + }, + { + "epoch": 1.4170431290783427, + "grad_norm": 1.35493004322052, + "learning_rate": 0.000105533032163404, + "loss": 0.1491, + "step": 36700 + }, + { + "epoch": 1.4174292443723697, + "grad_norm": 0.9102393984794617, + "learning_rate": 0.0001055072911438022, + "loss": 0.1565, + "step": 36710 + }, + { + "epoch": 1.4178153596663963, + "grad_norm": 1.7878345251083374, + "learning_rate": 0.00010548155012420043, + "loss": 0.4864, + "step": 36720 + }, + { + "epoch": 1.4182014749604233, + "grad_norm": 1.4333637952804565, + "learning_rate": 0.00010545580910459864, + "loss": 0.3246, + "step": 36730 + }, + { + "epoch": 1.41858759025445, + "grad_norm": 0.0966360941529274, + "learning_rate": 0.00010543006808499684, + "loss": 0.1758, + "step": 36740 + }, + { + "epoch": 1.4189737055484768, + "grad_norm": 0.30043545365333557, + "learning_rate": 0.00010540432706539507, + "loss": 0.1811, + "step": 36750 + }, + { + "epoch": 1.4193598208425036, + "grad_norm": 1.7705951929092407, + "learning_rate": 0.00010537858604579327, + "loss": 0.1824, + "step": 36760 + }, + { + "epoch": 1.4197459361365303, + "grad_norm": 1.2025195360183716, + "learning_rate": 0.0001053528450261915, + "loss": 0.1463, + "step": 36770 + }, + { + "epoch": 1.420132051430557, + "grad_norm": 0.3154304325580597, + "learning_rate": 0.00010532710400658972, + "loss": 0.1817, + "step": 36780 + }, + { + "epoch": 1.4205181667245839, + "grad_norm": 3.392331838607788, + "learning_rate": 0.00010530136298698792, + "loss": 0.2938, + "step": 36790 + }, + { + "epoch": 1.4209042820186109, + "grad_norm": 0.7256132364273071, + "learning_rate": 0.00010527562196738613, + "loss": 0.28, + "step": 36800 + }, + { + "epoch": 1.4212903973126376, + "grad_norm": 1.9007991552352905, + "learning_rate": 0.00010524988094778435, + "loss": 0.2306, + "step": 36810 + }, + { + "epoch": 1.4216765126066644, + "grad_norm": 3.21189546585083, + "learning_rate": 0.00010522413992818256, + "loss": 0.2945, + "step": 36820 + }, + { + "epoch": 1.4220626279006912, + "grad_norm": 0.20476600527763367, + "learning_rate": 0.00010519839890858076, + "loss": 0.3129, + "step": 36830 + }, + { + "epoch": 1.422448743194718, + "grad_norm": 1.4225107431411743, + "learning_rate": 0.00010517265788897899, + "loss": 0.2262, + "step": 36840 + }, + { + "epoch": 1.4228348584887447, + "grad_norm": 1.203728437423706, + "learning_rate": 0.00010514691686937721, + "loss": 0.176, + "step": 36850 + }, + { + "epoch": 1.4232209737827715, + "grad_norm": 0.3001759648323059, + "learning_rate": 0.00010512117584977541, + "loss": 0.218, + "step": 36860 + }, + { + "epoch": 1.4236070890767984, + "grad_norm": 2.3143389225006104, + "learning_rate": 0.00010509543483017363, + "loss": 0.1562, + "step": 36870 + }, + { + "epoch": 1.423993204370825, + "grad_norm": 0.5087364912033081, + "learning_rate": 0.00010506969381057184, + "loss": 0.1207, + "step": 36880 + }, + { + "epoch": 1.424379319664852, + "grad_norm": 1.6521960496902466, + "learning_rate": 0.00010504395279097005, + "loss": 0.4756, + "step": 36890 + }, + { + "epoch": 1.4247654349588788, + "grad_norm": 0.09236706793308258, + "learning_rate": 0.00010501821177136825, + "loss": 0.1755, + "step": 36900 + }, + { + "epoch": 1.4251515502529055, + "grad_norm": 0.2143094390630722, + "learning_rate": 0.00010499247075176648, + "loss": 0.2126, + "step": 36910 + }, + { + "epoch": 1.4255376655469323, + "grad_norm": 1.210170865058899, + "learning_rate": 0.00010496672973216471, + "loss": 0.1391, + "step": 36920 + }, + { + "epoch": 1.425923780840959, + "grad_norm": 0.6248244047164917, + "learning_rate": 0.00010494098871256291, + "loss": 0.2898, + "step": 36930 + }, + { + "epoch": 1.426309896134986, + "grad_norm": 2.1100337505340576, + "learning_rate": 0.00010491524769296113, + "loss": 0.1198, + "step": 36940 + }, + { + "epoch": 1.4266960114290126, + "grad_norm": 2.5673348903656006, + "learning_rate": 0.00010488950667335933, + "loss": 0.2284, + "step": 36950 + }, + { + "epoch": 1.4270821267230396, + "grad_norm": 1.868195652961731, + "learning_rate": 0.00010486376565375755, + "loss": 0.3738, + "step": 36960 + }, + { + "epoch": 1.4274682420170663, + "grad_norm": 1.0951671600341797, + "learning_rate": 0.00010483802463415577, + "loss": 0.1424, + "step": 36970 + }, + { + "epoch": 1.4278543573110931, + "grad_norm": 0.4791143536567688, + "learning_rate": 0.00010481228361455397, + "loss": 0.1445, + "step": 36980 + }, + { + "epoch": 1.4282404726051199, + "grad_norm": 3.2477540969848633, + "learning_rate": 0.0001047865425949522, + "loss": 0.2676, + "step": 36990 + }, + { + "epoch": 1.4286265878991466, + "grad_norm": 0.8082342147827148, + "learning_rate": 0.0001047608015753504, + "loss": 0.2295, + "step": 37000 + }, + { + "epoch": 1.4290127031931734, + "grad_norm": 2.3048954010009766, + "learning_rate": 0.00010473506055574863, + "loss": 0.1996, + "step": 37010 + }, + { + "epoch": 1.4293988184872002, + "grad_norm": 0.42648783326148987, + "learning_rate": 0.00010470931953614683, + "loss": 0.2549, + "step": 37020 + }, + { + "epoch": 1.4297849337812272, + "grad_norm": 1.8794362545013428, + "learning_rate": 0.00010468357851654504, + "loss": 0.1929, + "step": 37030 + }, + { + "epoch": 1.430171049075254, + "grad_norm": 0.9679039120674133, + "learning_rate": 0.00010465783749694327, + "loss": 0.1544, + "step": 37040 + }, + { + "epoch": 1.4305571643692807, + "grad_norm": 0.7789367437362671, + "learning_rate": 0.00010463209647734147, + "loss": 0.2251, + "step": 37050 + }, + { + "epoch": 1.4309432796633075, + "grad_norm": 1.1705437898635864, + "learning_rate": 0.0001046063554577397, + "loss": 0.1567, + "step": 37060 + }, + { + "epoch": 1.4313293949573342, + "grad_norm": 0.13087430596351624, + "learning_rate": 0.00010458061443813789, + "loss": 0.3237, + "step": 37070 + }, + { + "epoch": 1.431715510251361, + "grad_norm": 1.0658761262893677, + "learning_rate": 0.00010455487341853612, + "loss": 0.24, + "step": 37080 + }, + { + "epoch": 1.4321016255453878, + "grad_norm": 1.686922311782837, + "learning_rate": 0.00010452913239893433, + "loss": 0.1911, + "step": 37090 + }, + { + "epoch": 1.4324877408394148, + "grad_norm": 2.010221481323242, + "learning_rate": 0.00010450339137933253, + "loss": 0.4569, + "step": 37100 + }, + { + "epoch": 1.4328738561334413, + "grad_norm": 0.8007562160491943, + "learning_rate": 0.00010447765035973076, + "loss": 0.1198, + "step": 37110 + }, + { + "epoch": 1.4332599714274683, + "grad_norm": 0.5455211997032166, + "learning_rate": 0.00010445190934012896, + "loss": 0.2179, + "step": 37120 + }, + { + "epoch": 1.433646086721495, + "grad_norm": 1.4133542776107788, + "learning_rate": 0.00010442616832052719, + "loss": 0.45, + "step": 37130 + }, + { + "epoch": 1.4340322020155218, + "grad_norm": 1.6218222379684448, + "learning_rate": 0.00010440042730092539, + "loss": 0.19, + "step": 37140 + }, + { + "epoch": 1.4344183173095486, + "grad_norm": 0.6425970196723938, + "learning_rate": 0.00010437468628132361, + "loss": 0.2345, + "step": 37150 + }, + { + "epoch": 1.4348044326035754, + "grad_norm": 0.4344918131828308, + "learning_rate": 0.00010434894526172183, + "loss": 0.2532, + "step": 37160 + }, + { + "epoch": 1.4351905478976021, + "grad_norm": 0.6628998517990112, + "learning_rate": 0.00010432320424212003, + "loss": 0.2281, + "step": 37170 + }, + { + "epoch": 1.435576663191629, + "grad_norm": 0.8695842027664185, + "learning_rate": 0.00010429746322251825, + "loss": 0.2796, + "step": 37180 + }, + { + "epoch": 1.435962778485656, + "grad_norm": 0.16481854021549225, + "learning_rate": 0.00010427172220291645, + "loss": 0.2333, + "step": 37190 + }, + { + "epoch": 1.4363488937796827, + "grad_norm": 0.7194454073905945, + "learning_rate": 0.00010424598118331468, + "loss": 0.1413, + "step": 37200 + }, + { + "epoch": 1.4367350090737094, + "grad_norm": 4.845508575439453, + "learning_rate": 0.00010422024016371288, + "loss": 0.3944, + "step": 37210 + }, + { + "epoch": 1.4371211243677362, + "grad_norm": 2.6386618614196777, + "learning_rate": 0.0001041944991441111, + "loss": 0.3037, + "step": 37220 + }, + { + "epoch": 1.437507239661763, + "grad_norm": 0.4089922308921814, + "learning_rate": 0.00010416875812450932, + "loss": 0.2985, + "step": 37230 + }, + { + "epoch": 1.4378933549557897, + "grad_norm": 1.456944465637207, + "learning_rate": 0.00010414301710490752, + "loss": 0.3031, + "step": 37240 + }, + { + "epoch": 1.4382794702498165, + "grad_norm": 1.301829218864441, + "learning_rate": 0.00010411727608530575, + "loss": 0.2578, + "step": 37250 + }, + { + "epoch": 1.4386655855438435, + "grad_norm": 1.2072703838348389, + "learning_rate": 0.00010409153506570395, + "loss": 0.391, + "step": 37260 + }, + { + "epoch": 1.4390517008378703, + "grad_norm": 0.5538531541824341, + "learning_rate": 0.00010406579404610217, + "loss": 0.1867, + "step": 37270 + }, + { + "epoch": 1.439437816131897, + "grad_norm": 1.0898678302764893, + "learning_rate": 0.0001040400530265004, + "loss": 0.2112, + "step": 37280 + }, + { + "epoch": 1.4398239314259238, + "grad_norm": 1.5769239664077759, + "learning_rate": 0.0001040143120068986, + "loss": 0.3121, + "step": 37290 + }, + { + "epoch": 1.4402100467199506, + "grad_norm": 0.3964422345161438, + "learning_rate": 0.00010398857098729681, + "loss": 0.196, + "step": 37300 + }, + { + "epoch": 1.4405961620139773, + "grad_norm": 1.0268182754516602, + "learning_rate": 0.00010396282996769501, + "loss": 0.1785, + "step": 37310 + }, + { + "epoch": 1.440982277308004, + "grad_norm": 1.750826358795166, + "learning_rate": 0.00010393708894809324, + "loss": 0.3654, + "step": 37320 + }, + { + "epoch": 1.441368392602031, + "grad_norm": 1.1231745481491089, + "learning_rate": 0.00010391134792849144, + "loss": 0.2594, + "step": 37330 + }, + { + "epoch": 1.4417545078960576, + "grad_norm": 0.2897786498069763, + "learning_rate": 0.00010388560690888967, + "loss": 0.3483, + "step": 37340 + }, + { + "epoch": 1.4421406231900846, + "grad_norm": 0.07170752435922623, + "learning_rate": 0.00010385986588928789, + "loss": 0.1462, + "step": 37350 + }, + { + "epoch": 1.4425267384841114, + "grad_norm": 3.326099395751953, + "learning_rate": 0.00010383412486968609, + "loss": 0.3785, + "step": 37360 + }, + { + "epoch": 1.4429128537781382, + "grad_norm": 0.5063263773918152, + "learning_rate": 0.0001038083838500843, + "loss": 0.2974, + "step": 37370 + }, + { + "epoch": 1.443298969072165, + "grad_norm": 1.160088062286377, + "learning_rate": 0.00010378264283048252, + "loss": 0.3003, + "step": 37380 + }, + { + "epoch": 1.4436850843661917, + "grad_norm": 0.7043284177780151, + "learning_rate": 0.00010375690181088073, + "loss": 0.3193, + "step": 37390 + }, + { + "epoch": 1.4440711996602185, + "grad_norm": 0.2916620373725891, + "learning_rate": 0.00010373116079127896, + "loss": 0.235, + "step": 37400 + }, + { + "epoch": 1.4444573149542452, + "grad_norm": 2.1940219402313232, + "learning_rate": 0.00010370541977167716, + "loss": 0.2657, + "step": 37410 + }, + { + "epoch": 1.4448434302482722, + "grad_norm": 1.849794626235962, + "learning_rate": 0.00010367967875207539, + "loss": 0.3121, + "step": 37420 + }, + { + "epoch": 1.445229545542299, + "grad_norm": 1.1139589548110962, + "learning_rate": 0.00010365393773247359, + "loss": 0.2179, + "step": 37430 + }, + { + "epoch": 1.4456156608363258, + "grad_norm": 0.917142927646637, + "learning_rate": 0.0001036281967128718, + "loss": 0.2406, + "step": 37440 + }, + { + "epoch": 1.4460017761303525, + "grad_norm": 1.6973673105239868, + "learning_rate": 0.00010360245569327001, + "loss": 0.3236, + "step": 37450 + }, + { + "epoch": 1.4463878914243793, + "grad_norm": 0.3979933559894562, + "learning_rate": 0.00010357671467366823, + "loss": 0.1477, + "step": 37460 + }, + { + "epoch": 1.446774006718406, + "grad_norm": 0.8938451409339905, + "learning_rate": 0.00010355097365406645, + "loss": 0.116, + "step": 37470 + }, + { + "epoch": 1.4471601220124328, + "grad_norm": 1.068787693977356, + "learning_rate": 0.00010352523263446465, + "loss": 0.2342, + "step": 37480 + }, + { + "epoch": 1.4475462373064598, + "grad_norm": 1.1985591650009155, + "learning_rate": 0.00010349949161486288, + "loss": 0.2423, + "step": 37490 + }, + { + "epoch": 1.4479323526004866, + "grad_norm": 1.0383973121643066, + "learning_rate": 0.00010347375059526108, + "loss": 0.3059, + "step": 37500 + }, + { + "epoch": 1.4483184678945134, + "grad_norm": 0.26652297377586365, + "learning_rate": 0.0001034480095756593, + "loss": 0.262, + "step": 37510 + }, + { + "epoch": 1.4487045831885401, + "grad_norm": 1.8498083353042603, + "learning_rate": 0.0001034222685560575, + "loss": 0.2122, + "step": 37520 + }, + { + "epoch": 1.4490906984825669, + "grad_norm": 1.2896068096160889, + "learning_rate": 0.00010339652753645572, + "loss": 0.2939, + "step": 37530 + }, + { + "epoch": 1.4494768137765937, + "grad_norm": 1.423343300819397, + "learning_rate": 0.00010337078651685395, + "loss": 0.3375, + "step": 37540 + }, + { + "epoch": 1.4498629290706204, + "grad_norm": 1.8248246908187866, + "learning_rate": 0.00010334504549725215, + "loss": 0.3534, + "step": 37550 + }, + { + "epoch": 1.4502490443646474, + "grad_norm": 2.3713393211364746, + "learning_rate": 0.00010331930447765037, + "loss": 0.2795, + "step": 37560 + }, + { + "epoch": 1.450635159658674, + "grad_norm": 2.1431849002838135, + "learning_rate": 0.00010329356345804857, + "loss": 0.2503, + "step": 37570 + }, + { + "epoch": 1.451021274952701, + "grad_norm": 1.6521297693252563, + "learning_rate": 0.0001032678224384468, + "loss": 0.1789, + "step": 37580 + }, + { + "epoch": 1.4514073902467277, + "grad_norm": 1.5589754581451416, + "learning_rate": 0.00010324208141884501, + "loss": 0.3201, + "step": 37590 + }, + { + "epoch": 1.4517935055407545, + "grad_norm": 0.4339803159236908, + "learning_rate": 0.00010321634039924321, + "loss": 0.2964, + "step": 37600 + }, + { + "epoch": 1.4521796208347812, + "grad_norm": 0.30054792761802673, + "learning_rate": 0.00010319059937964144, + "loss": 0.3477, + "step": 37610 + }, + { + "epoch": 1.452565736128808, + "grad_norm": 2.2864038944244385, + "learning_rate": 0.00010316485836003964, + "loss": 0.2849, + "step": 37620 + }, + { + "epoch": 1.4529518514228348, + "grad_norm": 1.8392651081085205, + "learning_rate": 0.00010313911734043787, + "loss": 0.2089, + "step": 37630 + }, + { + "epoch": 1.4533379667168616, + "grad_norm": 1.0444347858428955, + "learning_rate": 0.00010311337632083607, + "loss": 0.2306, + "step": 37640 + }, + { + "epoch": 1.4537240820108885, + "grad_norm": 1.265647053718567, + "learning_rate": 0.00010308763530123429, + "loss": 0.2192, + "step": 37650 + }, + { + "epoch": 1.4541101973049153, + "grad_norm": 1.0464913845062256, + "learning_rate": 0.0001030618942816325, + "loss": 0.2857, + "step": 37660 + }, + { + "epoch": 1.454496312598942, + "grad_norm": 0.5931240320205688, + "learning_rate": 0.0001030361532620307, + "loss": 0.1751, + "step": 37670 + }, + { + "epoch": 1.4548824278929688, + "grad_norm": 2.0155346393585205, + "learning_rate": 0.00010301041224242893, + "loss": 0.4124, + "step": 37680 + }, + { + "epoch": 1.4552685431869956, + "grad_norm": 2.843345880508423, + "learning_rate": 0.00010298467122282713, + "loss": 0.2119, + "step": 37690 + }, + { + "epoch": 1.4556546584810224, + "grad_norm": 1.977612018585205, + "learning_rate": 0.00010295893020322536, + "loss": 0.271, + "step": 37700 + }, + { + "epoch": 1.4560407737750491, + "grad_norm": 1.2862237691879272, + "learning_rate": 0.00010293318918362356, + "loss": 0.4147, + "step": 37710 + }, + { + "epoch": 1.4564268890690761, + "grad_norm": 0.8235340118408203, + "learning_rate": 0.00010290744816402179, + "loss": 0.1918, + "step": 37720 + }, + { + "epoch": 1.456813004363103, + "grad_norm": 1.0351753234863281, + "learning_rate": 0.00010288170714442, + "loss": 0.2343, + "step": 37730 + }, + { + "epoch": 1.4571991196571297, + "grad_norm": 0.2567160725593567, + "learning_rate": 0.0001028559661248182, + "loss": 0.2664, + "step": 37740 + }, + { + "epoch": 1.4575852349511564, + "grad_norm": 2.7498669624328613, + "learning_rate": 0.00010283022510521643, + "loss": 0.2839, + "step": 37750 + }, + { + "epoch": 1.4579713502451832, + "grad_norm": 0.7125422358512878, + "learning_rate": 0.00010280448408561463, + "loss": 0.2411, + "step": 37760 + }, + { + "epoch": 1.45835746553921, + "grad_norm": 0.720761239528656, + "learning_rate": 0.00010277874306601285, + "loss": 0.315, + "step": 37770 + }, + { + "epoch": 1.4587435808332367, + "grad_norm": 3.188563346862793, + "learning_rate": 0.00010275300204641108, + "loss": 0.3945, + "step": 37780 + }, + { + "epoch": 1.4591296961272637, + "grad_norm": 1.23422110080719, + "learning_rate": 0.00010272726102680928, + "loss": 0.1828, + "step": 37790 + }, + { + "epoch": 1.4595158114212903, + "grad_norm": 1.1572456359863281, + "learning_rate": 0.00010270152000720749, + "loss": 0.1687, + "step": 37800 + }, + { + "epoch": 1.4599019267153173, + "grad_norm": 0.6565262079238892, + "learning_rate": 0.00010267577898760569, + "loss": 0.3145, + "step": 37810 + }, + { + "epoch": 1.460288042009344, + "grad_norm": 0.6239646673202515, + "learning_rate": 0.00010265003796800392, + "loss": 0.2381, + "step": 37820 + }, + { + "epoch": 1.4606741573033708, + "grad_norm": 0.7578912377357483, + "learning_rate": 0.00010262429694840212, + "loss": 0.2019, + "step": 37830 + }, + { + "epoch": 1.4610602725973976, + "grad_norm": 1.0978549718856812, + "learning_rate": 0.00010259855592880035, + "loss": 0.2477, + "step": 37840 + }, + { + "epoch": 1.4614463878914243, + "grad_norm": 0.3615519404411316, + "learning_rate": 0.00010257281490919857, + "loss": 0.4241, + "step": 37850 + }, + { + "epoch": 1.461832503185451, + "grad_norm": 0.10355047881603241, + "learning_rate": 0.00010254707388959677, + "loss": 0.1181, + "step": 37860 + }, + { + "epoch": 1.4622186184794779, + "grad_norm": 0.23957425355911255, + "learning_rate": 0.00010252133286999499, + "loss": 0.1675, + "step": 37870 + }, + { + "epoch": 1.4626047337735049, + "grad_norm": 1.7581062316894531, + "learning_rate": 0.00010249559185039318, + "loss": 0.241, + "step": 37880 + }, + { + "epoch": 1.4629908490675316, + "grad_norm": 1.660989761352539, + "learning_rate": 0.00010246985083079141, + "loss": 0.2099, + "step": 37890 + }, + { + "epoch": 1.4633769643615584, + "grad_norm": 1.1359142065048218, + "learning_rate": 0.00010244410981118964, + "loss": 0.3686, + "step": 37900 + }, + { + "epoch": 1.4637630796555852, + "grad_norm": 0.9489149451255798, + "learning_rate": 0.00010241836879158784, + "loss": 0.3616, + "step": 37910 + }, + { + "epoch": 1.464149194949612, + "grad_norm": 1.5209956169128418, + "learning_rate": 0.00010239262777198607, + "loss": 0.3159, + "step": 37920 + }, + { + "epoch": 1.4645353102436387, + "grad_norm": 0.848943293094635, + "learning_rate": 0.00010236688675238427, + "loss": 0.3228, + "step": 37930 + }, + { + "epoch": 1.4649214255376655, + "grad_norm": 1.1476777791976929, + "learning_rate": 0.00010234114573278248, + "loss": 0.1873, + "step": 37940 + }, + { + "epoch": 1.4653075408316925, + "grad_norm": 1.4093862771987915, + "learning_rate": 0.00010231540471318068, + "loss": 0.5554, + "step": 37950 + }, + { + "epoch": 1.465693656125719, + "grad_norm": 0.17349161207675934, + "learning_rate": 0.0001022896636935789, + "loss": 0.2992, + "step": 37960 + }, + { + "epoch": 1.466079771419746, + "grad_norm": 0.28885993361473083, + "learning_rate": 0.00010226392267397713, + "loss": 0.2618, + "step": 37970 + }, + { + "epoch": 1.4664658867137728, + "grad_norm": 1.1087830066680908, + "learning_rate": 0.00010223818165437533, + "loss": 0.2647, + "step": 37980 + }, + { + "epoch": 1.4668520020077995, + "grad_norm": 0.08338876068592072, + "learning_rate": 0.00010221244063477356, + "loss": 0.3057, + "step": 37990 + }, + { + "epoch": 1.4672381173018263, + "grad_norm": 2.159362316131592, + "learning_rate": 0.00010218669961517176, + "loss": 0.1499, + "step": 38000 + }, + { + "epoch": 1.467624232595853, + "grad_norm": 0.8207988142967224, + "learning_rate": 0.00010216095859556997, + "loss": 0.1482, + "step": 38010 + }, + { + "epoch": 1.46801034788988, + "grad_norm": 0.5458611845970154, + "learning_rate": 0.00010213521757596818, + "loss": 0.2794, + "step": 38020 + }, + { + "epoch": 1.4683964631839066, + "grad_norm": 1.6955047845840454, + "learning_rate": 0.0001021094765563664, + "loss": 0.2627, + "step": 38030 + }, + { + "epoch": 1.4687825784779336, + "grad_norm": 0.9796440601348877, + "learning_rate": 0.00010208373553676463, + "loss": 0.1924, + "step": 38040 + }, + { + "epoch": 1.4691686937719604, + "grad_norm": 0.9906508326530457, + "learning_rate": 0.00010205799451716282, + "loss": 0.1597, + "step": 38050 + }, + { + "epoch": 1.4695548090659871, + "grad_norm": 1.8590656518936157, + "learning_rate": 0.00010203225349756105, + "loss": 0.2874, + "step": 38060 + }, + { + "epoch": 1.469940924360014, + "grad_norm": 1.6403672695159912, + "learning_rate": 0.00010200651247795925, + "loss": 0.2926, + "step": 38070 + }, + { + "epoch": 1.4703270396540407, + "grad_norm": 0.4410895109176636, + "learning_rate": 0.00010198077145835746, + "loss": 0.1368, + "step": 38080 + }, + { + "epoch": 1.4707131549480674, + "grad_norm": 2.4955286979675293, + "learning_rate": 0.00010195503043875569, + "loss": 0.3195, + "step": 38090 + }, + { + "epoch": 1.4710992702420942, + "grad_norm": 1.1799029111862183, + "learning_rate": 0.00010192928941915389, + "loss": 0.1867, + "step": 38100 + }, + { + "epoch": 1.4714853855361212, + "grad_norm": 0.7959389090538025, + "learning_rate": 0.00010190354839955212, + "loss": 0.2988, + "step": 38110 + }, + { + "epoch": 1.471871500830148, + "grad_norm": 2.7750720977783203, + "learning_rate": 0.00010187780737995032, + "loss": 0.1886, + "step": 38120 + }, + { + "epoch": 1.4722576161241747, + "grad_norm": 1.5834373235702515, + "learning_rate": 0.00010185206636034854, + "loss": 0.3542, + "step": 38130 + }, + { + "epoch": 1.4726437314182015, + "grad_norm": 1.9757747650146484, + "learning_rate": 0.00010182632534074674, + "loss": 0.302, + "step": 38140 + }, + { + "epoch": 1.4730298467122283, + "grad_norm": 1.1752204895019531, + "learning_rate": 0.00010180058432114497, + "loss": 0.3535, + "step": 38150 + }, + { + "epoch": 1.473415962006255, + "grad_norm": 0.3877789378166199, + "learning_rate": 0.00010177484330154318, + "loss": 0.2508, + "step": 38160 + }, + { + "epoch": 1.4738020773002818, + "grad_norm": 0.13749545812606812, + "learning_rate": 0.00010174910228194138, + "loss": 0.2141, + "step": 38170 + }, + { + "epoch": 1.4741881925943088, + "grad_norm": 1.3663641214370728, + "learning_rate": 0.00010172336126233961, + "loss": 0.3231, + "step": 38180 + }, + { + "epoch": 1.4745743078883353, + "grad_norm": 1.6267393827438354, + "learning_rate": 0.00010169762024273781, + "loss": 0.3233, + "step": 38190 + }, + { + "epoch": 1.4749604231823623, + "grad_norm": 0.2993789315223694, + "learning_rate": 0.00010167187922313604, + "loss": 0.28, + "step": 38200 + }, + { + "epoch": 1.475346538476389, + "grad_norm": 0.16693222522735596, + "learning_rate": 0.00010164613820353424, + "loss": 0.188, + "step": 38210 + }, + { + "epoch": 1.4757326537704158, + "grad_norm": 0.6939979791641235, + "learning_rate": 0.00010162039718393246, + "loss": 0.263, + "step": 38220 + }, + { + "epoch": 1.4761187690644426, + "grad_norm": 0.37910985946655273, + "learning_rate": 0.00010159465616433068, + "loss": 0.1963, + "step": 38230 + }, + { + "epoch": 1.4765048843584694, + "grad_norm": 1.782188892364502, + "learning_rate": 0.00010156891514472888, + "loss": 0.3814, + "step": 38240 + }, + { + "epoch": 1.4768909996524964, + "grad_norm": 1.159278392791748, + "learning_rate": 0.0001015431741251271, + "loss": 0.2043, + "step": 38250 + }, + { + "epoch": 1.477277114946523, + "grad_norm": 1.09486985206604, + "learning_rate": 0.0001015174331055253, + "loss": 0.2128, + "step": 38260 + }, + { + "epoch": 1.47766323024055, + "grad_norm": 0.36655205488204956, + "learning_rate": 0.00010149169208592353, + "loss": 0.298, + "step": 38270 + }, + { + "epoch": 1.4780493455345767, + "grad_norm": 0.8908851742744446, + "learning_rate": 0.00010146595106632176, + "loss": 0.3707, + "step": 38280 + }, + { + "epoch": 1.4784354608286034, + "grad_norm": 0.251338928937912, + "learning_rate": 0.00010144021004671996, + "loss": 0.2495, + "step": 38290 + }, + { + "epoch": 1.4788215761226302, + "grad_norm": 1.0613712072372437, + "learning_rate": 0.00010141446902711817, + "loss": 0.2112, + "step": 38300 + }, + { + "epoch": 1.479207691416657, + "grad_norm": 1.459799885749817, + "learning_rate": 0.00010138872800751637, + "loss": 0.2595, + "step": 38310 + }, + { + "epoch": 1.4795938067106837, + "grad_norm": 2.6898603439331055, + "learning_rate": 0.0001013629869879146, + "loss": 0.2758, + "step": 38320 + }, + { + "epoch": 1.4799799220047105, + "grad_norm": 0.19628773629665375, + "learning_rate": 0.0001013372459683128, + "loss": 0.1843, + "step": 38330 + }, + { + "epoch": 1.4803660372987375, + "grad_norm": 2.0871078968048096, + "learning_rate": 0.00010131150494871102, + "loss": 0.1661, + "step": 38340 + }, + { + "epoch": 1.4807521525927643, + "grad_norm": 0.7689336538314819, + "learning_rate": 0.00010128576392910925, + "loss": 0.157, + "step": 38350 + }, + { + "epoch": 1.481138267886791, + "grad_norm": 1.4471644163131714, + "learning_rate": 0.00010126002290950745, + "loss": 0.2159, + "step": 38360 + }, + { + "epoch": 1.4815243831808178, + "grad_norm": 2.198559522628784, + "learning_rate": 0.00010123428188990566, + "loss": 0.5609, + "step": 38370 + }, + { + "epoch": 1.4819104984748446, + "grad_norm": 0.16012130677700043, + "learning_rate": 0.00010120854087030386, + "loss": 0.1979, + "step": 38380 + }, + { + "epoch": 1.4822966137688713, + "grad_norm": 1.0222225189208984, + "learning_rate": 0.00010118279985070209, + "loss": 0.1547, + "step": 38390 + }, + { + "epoch": 1.482682729062898, + "grad_norm": 2.7192416191101074, + "learning_rate": 0.00010115705883110032, + "loss": 0.4942, + "step": 38400 + }, + { + "epoch": 1.483068844356925, + "grad_norm": 1.887128472328186, + "learning_rate": 0.00010113131781149852, + "loss": 0.177, + "step": 38410 + }, + { + "epoch": 1.4834549596509516, + "grad_norm": 2.7628560066223145, + "learning_rate": 0.00010110557679189674, + "loss": 0.2931, + "step": 38420 + }, + { + "epoch": 1.4838410749449786, + "grad_norm": 0.3852572739124298, + "learning_rate": 0.00010107983577229494, + "loss": 0.3392, + "step": 38430 + }, + { + "epoch": 1.4842271902390054, + "grad_norm": 1.047448992729187, + "learning_rate": 0.00010105409475269316, + "loss": 0.3741, + "step": 38440 + }, + { + "epoch": 1.4846133055330322, + "grad_norm": 1.4930602312088013, + "learning_rate": 0.00010102835373309136, + "loss": 0.2564, + "step": 38450 + }, + { + "epoch": 1.484999420827059, + "grad_norm": 1.3012608289718628, + "learning_rate": 0.00010100261271348958, + "loss": 0.3376, + "step": 38460 + }, + { + "epoch": 1.4853855361210857, + "grad_norm": 2.163942337036133, + "learning_rate": 0.00010097687169388781, + "loss": 0.3548, + "step": 38470 + }, + { + "epoch": 1.4857716514151125, + "grad_norm": 1.864189624786377, + "learning_rate": 0.00010095113067428601, + "loss": 0.165, + "step": 38480 + }, + { + "epoch": 1.4861577667091392, + "grad_norm": 0.5661312341690063, + "learning_rate": 0.00010092538965468424, + "loss": 0.1764, + "step": 38490 + }, + { + "epoch": 1.4865438820031662, + "grad_norm": 0.13517481088638306, + "learning_rate": 0.00010089964863508244, + "loss": 0.5223, + "step": 38500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.344487878656e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..24bbab1de3826dffa96bf92e7b1d13fc034663eb --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35e83d3b002cb1d789981e91fa320dce6ab65241a59f683796e018cd904caf6e +size 292290560 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..25a37eb4126843b39a49a485bb4a5926bbaa12a9 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:967c4b983608b453cbc36aa67261e7bb217b1817c9e43be76ec9286b2bff8d03 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..0eb58ed46245776775a67beb5512699bbc22d868 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..d7b532659737416f7745fdb8c1ce5505e3243e96 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..69f7360eb0820759bf448ebe8b28b6e6dd872323 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/trainer_state.json @@ -0,0 +1,27334 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.505849646704506, + "eval_steps": 500, + "global_step": 39000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + }, + { + "epoch": 1.274566585582455, + "grad_norm": 0.5019121766090393, + "learning_rate": 0.0001150314683964632, + "loss": 0.1857, + "step": 33010 + }, + { + "epoch": 1.2749527008764816, + "grad_norm": 0.258952260017395, + "learning_rate": 0.0001150057273768614, + "loss": 0.11, + "step": 33020 + }, + { + "epoch": 1.2753388161705086, + "grad_norm": 0.8540348410606384, + "learning_rate": 0.00011497998635725962, + "loss": 0.1852, + "step": 33030 + }, + { + "epoch": 1.2757249314645354, + "grad_norm": 0.08689398318529129, + "learning_rate": 0.00011495424533765783, + "loss": 0.2225, + "step": 33040 + }, + { + "epoch": 1.2761110467585621, + "grad_norm": 0.42253822088241577, + "learning_rate": 0.00011492850431805605, + "loss": 0.0751, + "step": 33050 + }, + { + "epoch": 1.276497162052589, + "grad_norm": 1.2964017391204834, + "learning_rate": 0.00011490276329845426, + "loss": 0.2384, + "step": 33060 + }, + { + "epoch": 1.2768832773466157, + "grad_norm": 0.5337836146354675, + "learning_rate": 0.00011487702227885246, + "loss": 0.1415, + "step": 33070 + }, + { + "epoch": 1.2772693926406424, + "grad_norm": 2.7771682739257812, + "learning_rate": 0.00011485128125925069, + "loss": 0.378, + "step": 33080 + }, + { + "epoch": 1.2776555079346692, + "grad_norm": 1.5107232332229614, + "learning_rate": 0.00011482554023964889, + "loss": 0.2482, + "step": 33090 + }, + { + "epoch": 1.2780416232286962, + "grad_norm": 0.6885499358177185, + "learning_rate": 0.00011479979922004711, + "loss": 0.2122, + "step": 33100 + }, + { + "epoch": 1.278427738522723, + "grad_norm": 0.9016557335853577, + "learning_rate": 0.00011477405820044533, + "loss": 0.2841, + "step": 33110 + }, + { + "epoch": 1.2788138538167497, + "grad_norm": 1.9532525539398193, + "learning_rate": 0.00011474831718084354, + "loss": 0.2281, + "step": 33120 + }, + { + "epoch": 1.2791999691107765, + "grad_norm": 2.1078782081604004, + "learning_rate": 0.00011472257616124175, + "loss": 0.2836, + "step": 33130 + }, + { + "epoch": 1.2795860844048033, + "grad_norm": 0.19830390810966492, + "learning_rate": 0.00011469683514163995, + "loss": 0.388, + "step": 33140 + }, + { + "epoch": 1.27997219969883, + "grad_norm": 0.17538850009441376, + "learning_rate": 0.00011467109412203818, + "loss": 0.3274, + "step": 33150 + }, + { + "epoch": 1.2803583149928568, + "grad_norm": 0.7402139902114868, + "learning_rate": 0.00011464535310243641, + "loss": 0.1979, + "step": 33160 + }, + { + "epoch": 1.2807444302868838, + "grad_norm": 0.2097146362066269, + "learning_rate": 0.00011461961208283461, + "loss": 0.2464, + "step": 33170 + }, + { + "epoch": 1.2811305455809103, + "grad_norm": 1.2441083192825317, + "learning_rate": 0.00011459387106323283, + "loss": 0.266, + "step": 33180 + }, + { + "epoch": 1.2815166608749373, + "grad_norm": 2.518852710723877, + "learning_rate": 0.00011456813004363103, + "loss": 0.253, + "step": 33190 + }, + { + "epoch": 1.281902776168964, + "grad_norm": 0.8078998327255249, + "learning_rate": 0.00011454238902402925, + "loss": 0.2361, + "step": 33200 + }, + { + "epoch": 1.2822888914629909, + "grad_norm": 1.2297371625900269, + "learning_rate": 0.00011451664800442745, + "loss": 0.1974, + "step": 33210 + }, + { + "epoch": 1.2826750067570176, + "grad_norm": 0.4303855895996094, + "learning_rate": 0.00011449090698482567, + "loss": 0.3563, + "step": 33220 + }, + { + "epoch": 1.2830611220510444, + "grad_norm": 1.3215210437774658, + "learning_rate": 0.0001144651659652239, + "loss": 0.2818, + "step": 33230 + }, + { + "epoch": 1.2834472373450712, + "grad_norm": 1.546265959739685, + "learning_rate": 0.0001144394249456221, + "loss": 0.5778, + "step": 33240 + }, + { + "epoch": 1.283833352639098, + "grad_norm": 0.8895953297615051, + "learning_rate": 0.00011441368392602033, + "loss": 0.2485, + "step": 33250 + }, + { + "epoch": 1.284219467933125, + "grad_norm": 0.7534870505332947, + "learning_rate": 0.00011438794290641853, + "loss": 0.2825, + "step": 33260 + }, + { + "epoch": 1.2846055832271517, + "grad_norm": 0.052820973098278046, + "learning_rate": 0.00011436220188681674, + "loss": 0.2191, + "step": 33270 + }, + { + "epoch": 1.2849916985211784, + "grad_norm": 0.9264475107192993, + "learning_rate": 0.00011433646086721494, + "loss": 0.181, + "step": 33280 + }, + { + "epoch": 1.2853778138152052, + "grad_norm": 0.2128441333770752, + "learning_rate": 0.00011431071984761317, + "loss": 0.1819, + "step": 33290 + }, + { + "epoch": 1.285763929109232, + "grad_norm": 0.5400950312614441, + "learning_rate": 0.0001142849788280114, + "loss": 0.4316, + "step": 33300 + }, + { + "epoch": 1.2861500444032588, + "grad_norm": 0.8033271431922913, + "learning_rate": 0.00011425923780840959, + "loss": 0.2146, + "step": 33310 + }, + { + "epoch": 1.2865361596972855, + "grad_norm": 2.012575149536133, + "learning_rate": 0.00011423349678880782, + "loss": 0.4335, + "step": 33320 + }, + { + "epoch": 1.2869222749913125, + "grad_norm": 0.7352376580238342, + "learning_rate": 0.00011420775576920602, + "loss": 0.2124, + "step": 33330 + }, + { + "epoch": 1.2873083902853393, + "grad_norm": 0.769036591053009, + "learning_rate": 0.00011418201474960423, + "loss": 0.3602, + "step": 33340 + }, + { + "epoch": 1.287694505579366, + "grad_norm": 0.250592976808548, + "learning_rate": 0.00011415627373000246, + "loss": 0.1692, + "step": 33350 + }, + { + "epoch": 1.2880806208733928, + "grad_norm": 2.43820858001709, + "learning_rate": 0.00011413053271040066, + "loss": 0.2777, + "step": 33360 + }, + { + "epoch": 1.2884667361674196, + "grad_norm": 1.3179954290390015, + "learning_rate": 0.00011410479169079889, + "loss": 0.1794, + "step": 33370 + }, + { + "epoch": 1.2888528514614463, + "grad_norm": 1.0040466785430908, + "learning_rate": 0.00011407905067119709, + "loss": 0.3037, + "step": 33380 + }, + { + "epoch": 1.2892389667554731, + "grad_norm": 5.296288013458252, + "learning_rate": 0.00011405330965159531, + "loss": 0.2904, + "step": 33390 + }, + { + "epoch": 1.2896250820495, + "grad_norm": 0.4267273247241974, + "learning_rate": 0.00011402756863199351, + "loss": 0.2263, + "step": 33400 + }, + { + "epoch": 1.2900111973435266, + "grad_norm": 0.8817713260650635, + "learning_rate": 0.00011400182761239173, + "loss": 0.2715, + "step": 33410 + }, + { + "epoch": 1.2903973126375536, + "grad_norm": 2.7891275882720947, + "learning_rate": 0.00011397608659278995, + "loss": 0.2781, + "step": 33420 + }, + { + "epoch": 1.2907834279315804, + "grad_norm": 0.3013952374458313, + "learning_rate": 0.00011395034557318815, + "loss": 0.2563, + "step": 33430 + }, + { + "epoch": 1.2911695432256072, + "grad_norm": 1.766413927078247, + "learning_rate": 0.00011392460455358638, + "loss": 0.1854, + "step": 33440 + }, + { + "epoch": 1.291555658519634, + "grad_norm": 0.25331103801727295, + "learning_rate": 0.00011389886353398458, + "loss": 0.1206, + "step": 33450 + }, + { + "epoch": 1.2919417738136607, + "grad_norm": 0.036400288343429565, + "learning_rate": 0.0001138731225143828, + "loss": 0.4707, + "step": 33460 + }, + { + "epoch": 1.2923278891076875, + "grad_norm": 1.5620888471603394, + "learning_rate": 0.00011384738149478102, + "loss": 0.3312, + "step": 33470 + }, + { + "epoch": 1.2927140044017142, + "grad_norm": 0.6670392155647278, + "learning_rate": 0.00011382164047517922, + "loss": 0.2341, + "step": 33480 + }, + { + "epoch": 1.2931001196957412, + "grad_norm": 2.3108737468719482, + "learning_rate": 0.00011379589945557745, + "loss": 0.3843, + "step": 33490 + }, + { + "epoch": 1.293486234989768, + "grad_norm": 0.8025147318840027, + "learning_rate": 0.00011377015843597565, + "loss": 0.1982, + "step": 33500 + }, + { + "epoch": 1.2938723502837948, + "grad_norm": 1.7835719585418701, + "learning_rate": 0.00011374441741637387, + "loss": 0.3285, + "step": 33510 + }, + { + "epoch": 1.2942584655778215, + "grad_norm": 2.041508913040161, + "learning_rate": 0.00011371867639677207, + "loss": 0.2044, + "step": 33520 + }, + { + "epoch": 1.2946445808718483, + "grad_norm": 1.103378415107727, + "learning_rate": 0.0001136929353771703, + "loss": 0.1682, + "step": 33530 + }, + { + "epoch": 1.295030696165875, + "grad_norm": 0.057376351207494736, + "learning_rate": 0.00011366719435756851, + "loss": 0.1642, + "step": 33540 + }, + { + "epoch": 1.2954168114599018, + "grad_norm": 0.6539410948753357, + "learning_rate": 0.00011364145333796671, + "loss": 0.1549, + "step": 33550 + }, + { + "epoch": 1.2958029267539288, + "grad_norm": 1.250543236732483, + "learning_rate": 0.00011361571231836494, + "loss": 0.3764, + "step": 33560 + }, + { + "epoch": 1.2961890420479556, + "grad_norm": 0.23697887361049652, + "learning_rate": 0.00011358997129876314, + "loss": 0.3999, + "step": 33570 + }, + { + "epoch": 1.2965751573419824, + "grad_norm": 0.9318505525588989, + "learning_rate": 0.00011356423027916137, + "loss": 0.4156, + "step": 33580 + }, + { + "epoch": 1.2969612726360091, + "grad_norm": 1.3910777568817139, + "learning_rate": 0.00011353848925955957, + "loss": 0.3455, + "step": 33590 + }, + { + "epoch": 1.297347387930036, + "grad_norm": 1.6764451265335083, + "learning_rate": 0.00011351274823995779, + "loss": 0.1884, + "step": 33600 + }, + { + "epoch": 1.2977335032240627, + "grad_norm": 0.9300051927566528, + "learning_rate": 0.000113487007220356, + "loss": 0.119, + "step": 33610 + }, + { + "epoch": 1.2981196185180894, + "grad_norm": 2.447462558746338, + "learning_rate": 0.00011346126620075422, + "loss": 0.4403, + "step": 33620 + }, + { + "epoch": 1.2985057338121164, + "grad_norm": 1.216407060623169, + "learning_rate": 0.00011343552518115243, + "loss": 0.2415, + "step": 33630 + }, + { + "epoch": 1.298891849106143, + "grad_norm": 2.968648910522461, + "learning_rate": 0.00011340978416155063, + "loss": 0.2899, + "step": 33640 + }, + { + "epoch": 1.29927796440017, + "grad_norm": 0.6649970412254333, + "learning_rate": 0.00011338404314194886, + "loss": 0.3809, + "step": 33650 + }, + { + "epoch": 1.2996640796941967, + "grad_norm": 1.7277917861938477, + "learning_rate": 0.00011335830212234709, + "loss": 0.3308, + "step": 33660 + }, + { + "epoch": 1.3000501949882235, + "grad_norm": 1.3269709348678589, + "learning_rate": 0.00011333256110274529, + "loss": 0.3682, + "step": 33670 + }, + { + "epoch": 1.3004363102822503, + "grad_norm": 0.20609407126903534, + "learning_rate": 0.0001133068200831435, + "loss": 0.1379, + "step": 33680 + }, + { + "epoch": 1.300822425576277, + "grad_norm": 0.6592215299606323, + "learning_rate": 0.00011328107906354171, + "loss": 0.2746, + "step": 33690 + }, + { + "epoch": 1.3012085408703038, + "grad_norm": 1.903635859489441, + "learning_rate": 0.00011325533804393993, + "loss": 0.4729, + "step": 33700 + }, + { + "epoch": 1.3015946561643306, + "grad_norm": 0.8432504534721375, + "learning_rate": 0.00011322959702433813, + "loss": 0.2835, + "step": 33710 + }, + { + "epoch": 1.3019807714583576, + "grad_norm": 0.9862542152404785, + "learning_rate": 0.00011320385600473635, + "loss": 0.1823, + "step": 33720 + }, + { + "epoch": 1.3023668867523843, + "grad_norm": 3.845738649368286, + "learning_rate": 0.00011317811498513458, + "loss": 0.2542, + "step": 33730 + }, + { + "epoch": 1.302753002046411, + "grad_norm": 0.6317747235298157, + "learning_rate": 0.00011315237396553278, + "loss": 0.22, + "step": 33740 + }, + { + "epoch": 1.3031391173404379, + "grad_norm": 2.5221354961395264, + "learning_rate": 0.000113126632945931, + "loss": 0.2253, + "step": 33750 + }, + { + "epoch": 1.3035252326344646, + "grad_norm": 1.3326247930526733, + "learning_rate": 0.0001131008919263292, + "loss": 0.2021, + "step": 33760 + }, + { + "epoch": 1.3039113479284914, + "grad_norm": 1.148047685623169, + "learning_rate": 0.00011307515090672742, + "loss": 0.3987, + "step": 33770 + }, + { + "epoch": 1.3042974632225182, + "grad_norm": 0.19721268117427826, + "learning_rate": 0.00011304940988712562, + "loss": 0.2642, + "step": 33780 + }, + { + "epoch": 1.3046835785165452, + "grad_norm": 1.4060617685317993, + "learning_rate": 0.00011302366886752385, + "loss": 0.2736, + "step": 33790 + }, + { + "epoch": 1.305069693810572, + "grad_norm": 1.0736548900604248, + "learning_rate": 0.00011299792784792207, + "loss": 0.2448, + "step": 33800 + }, + { + "epoch": 1.3054558091045987, + "grad_norm": 4.352476119995117, + "learning_rate": 0.00011297218682832027, + "loss": 0.383, + "step": 33810 + }, + { + "epoch": 1.3058419243986255, + "grad_norm": 0.2249228060245514, + "learning_rate": 0.0001129464458087185, + "loss": 0.14, + "step": 33820 + }, + { + "epoch": 1.3062280396926522, + "grad_norm": 0.4820781946182251, + "learning_rate": 0.0001129207047891167, + "loss": 0.248, + "step": 33830 + }, + { + "epoch": 1.306614154986679, + "grad_norm": 2.2983391284942627, + "learning_rate": 0.00011289496376951491, + "loss": 0.2608, + "step": 33840 + }, + { + "epoch": 1.3070002702807058, + "grad_norm": 1.3315671682357788, + "learning_rate": 0.00011286922274991314, + "loss": 0.1727, + "step": 33850 + }, + { + "epoch": 1.3073863855747327, + "grad_norm": 2.060299873352051, + "learning_rate": 0.00011284348173031134, + "loss": 0.3097, + "step": 33860 + }, + { + "epoch": 1.3077725008687593, + "grad_norm": 2.096285581588745, + "learning_rate": 0.00011281774071070957, + "loss": 0.2305, + "step": 33870 + }, + { + "epoch": 1.3081586161627863, + "grad_norm": 0.4997636675834656, + "learning_rate": 0.00011279199969110777, + "loss": 0.1993, + "step": 33880 + }, + { + "epoch": 1.308544731456813, + "grad_norm": 0.58636474609375, + "learning_rate": 0.00011276625867150599, + "loss": 0.1873, + "step": 33890 + }, + { + "epoch": 1.3089308467508398, + "grad_norm": 0.9128592610359192, + "learning_rate": 0.00011274051765190419, + "loss": 0.1885, + "step": 33900 + }, + { + "epoch": 1.3093169620448666, + "grad_norm": 2.228043794631958, + "learning_rate": 0.0001127147766323024, + "loss": 0.3649, + "step": 33910 + }, + { + "epoch": 1.3097030773388934, + "grad_norm": 1.069002389907837, + "learning_rate": 0.00011268903561270063, + "loss": 0.5454, + "step": 33920 + }, + { + "epoch": 1.3100891926329201, + "grad_norm": 0.6207597851753235, + "learning_rate": 0.00011266329459309883, + "loss": 0.2329, + "step": 33930 + }, + { + "epoch": 1.3104753079269469, + "grad_norm": 1.262247920036316, + "learning_rate": 0.00011263755357349706, + "loss": 0.3437, + "step": 33940 + }, + { + "epoch": 1.3108614232209739, + "grad_norm": 1.7429994344711304, + "learning_rate": 0.00011261181255389526, + "loss": 0.228, + "step": 33950 + }, + { + "epoch": 1.3112475385150006, + "grad_norm": 0.646900475025177, + "learning_rate": 0.00011258607153429349, + "loss": 0.3739, + "step": 33960 + }, + { + "epoch": 1.3116336538090274, + "grad_norm": 1.8228782415390015, + "learning_rate": 0.0001125603305146917, + "loss": 0.2325, + "step": 33970 + }, + { + "epoch": 1.3120197691030542, + "grad_norm": 3.539228916168213, + "learning_rate": 0.0001125345894950899, + "loss": 0.194, + "step": 33980 + }, + { + "epoch": 1.312405884397081, + "grad_norm": 1.2801135778427124, + "learning_rate": 0.00011250884847548813, + "loss": 0.3069, + "step": 33990 + }, + { + "epoch": 1.3127919996911077, + "grad_norm": 3.6265695095062256, + "learning_rate": 0.00011248310745588633, + "loss": 0.3113, + "step": 34000 + }, + { + "epoch": 1.3131781149851345, + "grad_norm": 0.07370063662528992, + "learning_rate": 0.00011245736643628455, + "loss": 0.1449, + "step": 34010 + }, + { + "epoch": 1.3135642302791615, + "grad_norm": 1.0295637845993042, + "learning_rate": 0.00011243162541668275, + "loss": 0.22, + "step": 34020 + }, + { + "epoch": 1.313950345573188, + "grad_norm": 0.8803662061691284, + "learning_rate": 0.00011240588439708098, + "loss": 0.1368, + "step": 34030 + }, + { + "epoch": 1.314336460867215, + "grad_norm": 1.6597707271575928, + "learning_rate": 0.00011238014337747919, + "loss": 0.3038, + "step": 34040 + }, + { + "epoch": 1.3147225761612418, + "grad_norm": 2.115492343902588, + "learning_rate": 0.00011235440235787739, + "loss": 0.1754, + "step": 34050 + }, + { + "epoch": 1.3151086914552685, + "grad_norm": 0.8143919706344604, + "learning_rate": 0.00011232866133827562, + "loss": 0.3764, + "step": 34060 + }, + { + "epoch": 1.3154948067492953, + "grad_norm": 0.14369767904281616, + "learning_rate": 0.00011230292031867382, + "loss": 0.1742, + "step": 34070 + }, + { + "epoch": 1.315880922043322, + "grad_norm": 1.0129845142364502, + "learning_rate": 0.00011227717929907205, + "loss": 0.1458, + "step": 34080 + }, + { + "epoch": 1.316267037337349, + "grad_norm": 2.7300291061401367, + "learning_rate": 0.00011225143827947025, + "loss": 0.3939, + "step": 34090 + }, + { + "epoch": 1.3166531526313756, + "grad_norm": 0.20205609500408173, + "learning_rate": 0.00011222569725986847, + "loss": 0.29, + "step": 34100 + }, + { + "epoch": 1.3170392679254026, + "grad_norm": 1.8928464651107788, + "learning_rate": 0.00011219995624026669, + "loss": 0.1742, + "step": 34110 + }, + { + "epoch": 1.3174253832194294, + "grad_norm": 0.2639687955379486, + "learning_rate": 0.00011217421522066488, + "loss": 0.1745, + "step": 34120 + }, + { + "epoch": 1.3178114985134561, + "grad_norm": 0.5906389355659485, + "learning_rate": 0.00011214847420106311, + "loss": 0.2134, + "step": 34130 + }, + { + "epoch": 1.318197613807483, + "grad_norm": 0.9190629720687866, + "learning_rate": 0.00011212273318146131, + "loss": 0.2547, + "step": 34140 + }, + { + "epoch": 1.3185837291015097, + "grad_norm": 0.5749151110649109, + "learning_rate": 0.00011209699216185954, + "loss": 0.1688, + "step": 34150 + }, + { + "epoch": 1.3189698443955364, + "grad_norm": 0.82295823097229, + "learning_rate": 0.00011207125114225777, + "loss": 0.2884, + "step": 34160 + }, + { + "epoch": 1.3193559596895632, + "grad_norm": 0.07816460728645325, + "learning_rate": 0.00011204551012265597, + "loss": 0.2418, + "step": 34170 + }, + { + "epoch": 1.3197420749835902, + "grad_norm": 0.6417407393455505, + "learning_rate": 0.00011201976910305418, + "loss": 0.2557, + "step": 34180 + }, + { + "epoch": 1.320128190277617, + "grad_norm": 6.093267440795898, + "learning_rate": 0.00011199402808345238, + "loss": 0.3088, + "step": 34190 + }, + { + "epoch": 1.3205143055716437, + "grad_norm": 1.8861887454986572, + "learning_rate": 0.0001119682870638506, + "loss": 0.2204, + "step": 34200 + }, + { + "epoch": 1.3209004208656705, + "grad_norm": 2.3272714614868164, + "learning_rate": 0.0001119425460442488, + "loss": 0.2236, + "step": 34210 + }, + { + "epoch": 1.3212865361596973, + "grad_norm": 0.9608810544013977, + "learning_rate": 0.00011191680502464703, + "loss": 0.1897, + "step": 34220 + }, + { + "epoch": 1.321672651453724, + "grad_norm": 1.2157350778579712, + "learning_rate": 0.00011189106400504526, + "loss": 0.1526, + "step": 34230 + }, + { + "epoch": 1.3220587667477508, + "grad_norm": 1.6684671640396118, + "learning_rate": 0.00011186532298544346, + "loss": 0.3394, + "step": 34240 + }, + { + "epoch": 1.3224448820417778, + "grad_norm": 2.0432374477386475, + "learning_rate": 0.00011183958196584167, + "loss": 0.2183, + "step": 34250 + }, + { + "epoch": 1.3228309973358043, + "grad_norm": 0.9436892867088318, + "learning_rate": 0.00011181384094623988, + "loss": 0.2947, + "step": 34260 + }, + { + "epoch": 1.3232171126298313, + "grad_norm": 0.23260092735290527, + "learning_rate": 0.0001117880999266381, + "loss": 0.114, + "step": 34270 + }, + { + "epoch": 1.323603227923858, + "grad_norm": 1.2291594743728638, + "learning_rate": 0.0001117623589070363, + "loss": 0.3145, + "step": 34280 + }, + { + "epoch": 1.3239893432178849, + "grad_norm": 0.41411107778549194, + "learning_rate": 0.00011173661788743452, + "loss": 0.2937, + "step": 34290 + }, + { + "epoch": 1.3243754585119116, + "grad_norm": 2.354405164718628, + "learning_rate": 0.00011171087686783275, + "loss": 0.3933, + "step": 34300 + }, + { + "epoch": 1.3247615738059384, + "grad_norm": 2.6997978687286377, + "learning_rate": 0.00011168513584823095, + "loss": 0.1494, + "step": 34310 + }, + { + "epoch": 1.3251476890999654, + "grad_norm": 2.8430919647216797, + "learning_rate": 0.00011165939482862916, + "loss": 0.2869, + "step": 34320 + }, + { + "epoch": 1.325533804393992, + "grad_norm": 1.1737356185913086, + "learning_rate": 0.00011163365380902738, + "loss": 0.2792, + "step": 34330 + }, + { + "epoch": 1.325919919688019, + "grad_norm": 4.123973846435547, + "learning_rate": 0.00011160791278942559, + "loss": 0.5211, + "step": 34340 + }, + { + "epoch": 1.3263060349820457, + "grad_norm": 0.8862038850784302, + "learning_rate": 0.00011158217176982382, + "loss": 0.2976, + "step": 34350 + }, + { + "epoch": 1.3266921502760725, + "grad_norm": 1.8690590858459473, + "learning_rate": 0.00011155643075022202, + "loss": 0.2485, + "step": 34360 + }, + { + "epoch": 1.3270782655700992, + "grad_norm": 2.885589599609375, + "learning_rate": 0.00011153068973062024, + "loss": 0.3084, + "step": 34370 + }, + { + "epoch": 1.327464380864126, + "grad_norm": 0.9898788928985596, + "learning_rate": 0.00011150494871101844, + "loss": 0.261, + "step": 34380 + }, + { + "epoch": 1.3278504961581528, + "grad_norm": 0.6879653930664062, + "learning_rate": 0.00011147920769141667, + "loss": 0.2082, + "step": 34390 + }, + { + "epoch": 1.3282366114521795, + "grad_norm": 1.2619003057479858, + "learning_rate": 0.00011145346667181487, + "loss": 0.2402, + "step": 34400 + }, + { + "epoch": 1.3286227267462065, + "grad_norm": 1.1212007999420166, + "learning_rate": 0.00011142772565221308, + "loss": 0.3525, + "step": 34410 + }, + { + "epoch": 1.3290088420402333, + "grad_norm": 1.8431956768035889, + "learning_rate": 0.00011140198463261131, + "loss": 0.2212, + "step": 34420 + }, + { + "epoch": 1.32939495733426, + "grad_norm": 0.6185423731803894, + "learning_rate": 0.00011137624361300951, + "loss": 0.2455, + "step": 34430 + }, + { + "epoch": 1.3297810726282868, + "grad_norm": 2.3791301250457764, + "learning_rate": 0.00011135050259340774, + "loss": 0.1763, + "step": 34440 + }, + { + "epoch": 1.3301671879223136, + "grad_norm": 0.4928603768348694, + "learning_rate": 0.00011132476157380594, + "loss": 0.2381, + "step": 34450 + }, + { + "epoch": 1.3305533032163404, + "grad_norm": 1.5636029243469238, + "learning_rate": 0.00011129902055420416, + "loss": 0.1368, + "step": 34460 + }, + { + "epoch": 1.3309394185103671, + "grad_norm": 0.9425283074378967, + "learning_rate": 0.00011127327953460238, + "loss": 0.2825, + "step": 34470 + }, + { + "epoch": 1.3313255338043941, + "grad_norm": 1.2257115840911865, + "learning_rate": 0.00011124753851500058, + "loss": 0.2547, + "step": 34480 + }, + { + "epoch": 1.3317116490984207, + "grad_norm": 0.9416170716285706, + "learning_rate": 0.0001112217974953988, + "loss": 0.2766, + "step": 34490 + }, + { + "epoch": 1.3320977643924476, + "grad_norm": 0.5123847126960754, + "learning_rate": 0.000111196056475797, + "loss": 0.4733, + "step": 34500 + }, + { + "epoch": 1.3324838796864744, + "grad_norm": 1.5581384897232056, + "learning_rate": 0.00011117031545619523, + "loss": 0.1597, + "step": 34510 + }, + { + "epoch": 1.3328699949805012, + "grad_norm": 2.377333879470825, + "learning_rate": 0.00011114457443659343, + "loss": 0.209, + "step": 34520 + }, + { + "epoch": 1.333256110274528, + "grad_norm": 1.7840913534164429, + "learning_rate": 0.00011111883341699166, + "loss": 0.1759, + "step": 34530 + }, + { + "epoch": 1.3336422255685547, + "grad_norm": 1.1825993061065674, + "learning_rate": 0.00011109309239738987, + "loss": 0.2464, + "step": 34540 + }, + { + "epoch": 1.3340283408625815, + "grad_norm": 1.8859659433364868, + "learning_rate": 0.00011106735137778807, + "loss": 0.3539, + "step": 34550 + }, + { + "epoch": 1.3344144561566083, + "grad_norm": 1.9698175191879272, + "learning_rate": 0.0001110416103581863, + "loss": 0.3301, + "step": 34560 + }, + { + "epoch": 1.3348005714506352, + "grad_norm": 0.7649385333061218, + "learning_rate": 0.0001110158693385845, + "loss": 0.232, + "step": 34570 + }, + { + "epoch": 1.335186686744662, + "grad_norm": 0.56386399269104, + "learning_rate": 0.00011099012831898272, + "loss": 0.3425, + "step": 34580 + }, + { + "epoch": 1.3355728020386888, + "grad_norm": 2.956003189086914, + "learning_rate": 0.00011096438729938092, + "loss": 0.1518, + "step": 34590 + }, + { + "epoch": 1.3359589173327155, + "grad_norm": 2.612029552459717, + "learning_rate": 0.00011093864627977915, + "loss": 0.2765, + "step": 34600 + }, + { + "epoch": 1.3363450326267423, + "grad_norm": 0.9674397706985474, + "learning_rate": 0.00011091290526017736, + "loss": 0.303, + "step": 34610 + }, + { + "epoch": 1.336731147920769, + "grad_norm": 0.9578921794891357, + "learning_rate": 0.00011088716424057556, + "loss": 0.1405, + "step": 34620 + }, + { + "epoch": 1.3371172632147958, + "grad_norm": 2.168065071105957, + "learning_rate": 0.00011086142322097379, + "loss": 0.1914, + "step": 34630 + }, + { + "epoch": 1.3375033785088228, + "grad_norm": 1.3166526556015015, + "learning_rate": 0.00011083568220137199, + "loss": 0.4134, + "step": 34640 + }, + { + "epoch": 1.3378894938028496, + "grad_norm": 0.9082283973693848, + "learning_rate": 0.00011080994118177022, + "loss": 0.2693, + "step": 34650 + }, + { + "epoch": 1.3382756090968764, + "grad_norm": 2.203007698059082, + "learning_rate": 0.00011078420016216844, + "loss": 0.1847, + "step": 34660 + }, + { + "epoch": 1.3386617243909031, + "grad_norm": 0.8101674914360046, + "learning_rate": 0.00011075845914256664, + "loss": 0.3111, + "step": 34670 + }, + { + "epoch": 1.33904783968493, + "grad_norm": 1.9545695781707764, + "learning_rate": 0.00011073271812296486, + "loss": 0.3761, + "step": 34680 + }, + { + "epoch": 1.3394339549789567, + "grad_norm": 1.547581672668457, + "learning_rate": 0.00011070697710336306, + "loss": 0.2374, + "step": 34690 + }, + { + "epoch": 1.3398200702729834, + "grad_norm": 3.3519034385681152, + "learning_rate": 0.00011068123608376128, + "loss": 0.1957, + "step": 34700 + }, + { + "epoch": 1.3402061855670104, + "grad_norm": 1.5508599281311035, + "learning_rate": 0.00011065549506415948, + "loss": 0.4171, + "step": 34710 + }, + { + "epoch": 1.340592300861037, + "grad_norm": 1.8547546863555908, + "learning_rate": 0.00011062975404455771, + "loss": 0.1872, + "step": 34720 + }, + { + "epoch": 1.340978416155064, + "grad_norm": 1.4600756168365479, + "learning_rate": 0.00011060401302495594, + "loss": 0.3515, + "step": 34730 + }, + { + "epoch": 1.3413645314490907, + "grad_norm": 0.05774044618010521, + "learning_rate": 0.00011057827200535414, + "loss": 0.1604, + "step": 34740 + }, + { + "epoch": 1.3417506467431175, + "grad_norm": 2.8793342113494873, + "learning_rate": 0.00011055253098575235, + "loss": 0.3095, + "step": 34750 + }, + { + "epoch": 1.3421367620371443, + "grad_norm": 2.241042375564575, + "learning_rate": 0.00011052678996615055, + "loss": 0.2511, + "step": 34760 + }, + { + "epoch": 1.342522877331171, + "grad_norm": 1.9320632219314575, + "learning_rate": 0.00011050104894654878, + "loss": 0.4493, + "step": 34770 + }, + { + "epoch": 1.3429089926251978, + "grad_norm": 1.6483882665634155, + "learning_rate": 0.000110475307926947, + "loss": 0.217, + "step": 34780 + }, + { + "epoch": 1.3432951079192246, + "grad_norm": 0.9635765552520752, + "learning_rate": 0.0001104495669073452, + "loss": 0.5458, + "step": 34790 + }, + { + "epoch": 1.3436812232132516, + "grad_norm": 1.2436567544937134, + "learning_rate": 0.00011042382588774343, + "loss": 0.2857, + "step": 34800 + }, + { + "epoch": 1.3440673385072783, + "grad_norm": 2.8082425594329834, + "learning_rate": 0.00011039808486814163, + "loss": 0.3439, + "step": 34810 + }, + { + "epoch": 1.344453453801305, + "grad_norm": 1.0430901050567627, + "learning_rate": 0.00011037234384853984, + "loss": 0.1404, + "step": 34820 + }, + { + "epoch": 1.3448395690953319, + "grad_norm": 1.7387149333953857, + "learning_rate": 0.00011034660282893806, + "loss": 0.395, + "step": 34830 + }, + { + "epoch": 1.3452256843893586, + "grad_norm": 1.2713748216629028, + "learning_rate": 0.00011032086180933627, + "loss": 0.29, + "step": 34840 + }, + { + "epoch": 1.3456117996833854, + "grad_norm": 0.26068204641342163, + "learning_rate": 0.0001102951207897345, + "loss": 0.1814, + "step": 34850 + }, + { + "epoch": 1.3459979149774122, + "grad_norm": 2.4163243770599365, + "learning_rate": 0.0001102693797701327, + "loss": 0.2445, + "step": 34860 + }, + { + "epoch": 1.3463840302714392, + "grad_norm": 2.2439687252044678, + "learning_rate": 0.00011024363875053092, + "loss": 0.3338, + "step": 34870 + }, + { + "epoch": 1.346770145565466, + "grad_norm": 0.2822403609752655, + "learning_rate": 0.00011021789773092912, + "loss": 0.1648, + "step": 34880 + }, + { + "epoch": 1.3471562608594927, + "grad_norm": 0.07319017499685287, + "learning_rate": 0.00011019215671132734, + "loss": 0.107, + "step": 34890 + }, + { + "epoch": 1.3475423761535195, + "grad_norm": 0.9809044003486633, + "learning_rate": 0.00011016641569172555, + "loss": 0.256, + "step": 34900 + }, + { + "epoch": 1.3479284914475462, + "grad_norm": 0.5016226768493652, + "learning_rate": 0.00011014067467212376, + "loss": 0.3229, + "step": 34910 + }, + { + "epoch": 1.348314606741573, + "grad_norm": 1.3026005029678345, + "learning_rate": 0.00011011493365252199, + "loss": 0.2145, + "step": 34920 + }, + { + "epoch": 1.3487007220355998, + "grad_norm": 1.0752215385437012, + "learning_rate": 0.00011008919263292019, + "loss": 0.2355, + "step": 34930 + }, + { + "epoch": 1.3490868373296268, + "grad_norm": 2.2703003883361816, + "learning_rate": 0.00011006345161331842, + "loss": 0.2079, + "step": 34940 + }, + { + "epoch": 1.3494729526236533, + "grad_norm": 1.1323810815811157, + "learning_rate": 0.00011003771059371662, + "loss": 0.1015, + "step": 34950 + }, + { + "epoch": 1.3498590679176803, + "grad_norm": 0.10813555121421814, + "learning_rate": 0.00011001196957411484, + "loss": 0.4214, + "step": 34960 + }, + { + "epoch": 1.350245183211707, + "grad_norm": 0.07815568149089813, + "learning_rate": 0.00010998622855451306, + "loss": 0.1211, + "step": 34970 + }, + { + "epoch": 1.3506312985057338, + "grad_norm": 0.6748234629631042, + "learning_rate": 0.00010996048753491126, + "loss": 0.3508, + "step": 34980 + }, + { + "epoch": 1.3510174137997606, + "grad_norm": 1.8556997776031494, + "learning_rate": 0.00010993474651530948, + "loss": 0.2268, + "step": 34990 + }, + { + "epoch": 1.3514035290937874, + "grad_norm": 0.8696061372756958, + "learning_rate": 0.00010990900549570768, + "loss": 0.4321, + "step": 35000 + }, + { + "epoch": 1.3517896443878141, + "grad_norm": 0.42442765831947327, + "learning_rate": 0.00010988326447610591, + "loss": 0.1944, + "step": 35010 + }, + { + "epoch": 1.352175759681841, + "grad_norm": 1.0474554300308228, + "learning_rate": 0.00010985752345650411, + "loss": 0.1342, + "step": 35020 + }, + { + "epoch": 1.3525618749758679, + "grad_norm": 0.607037365436554, + "learning_rate": 0.00010983178243690234, + "loss": 0.2965, + "step": 35030 + }, + { + "epoch": 1.3529479902698947, + "grad_norm": 1.8160990476608276, + "learning_rate": 0.00010980604141730055, + "loss": 0.3192, + "step": 35040 + }, + { + "epoch": 1.3533341055639214, + "grad_norm": 2.0026509761810303, + "learning_rate": 0.00010978030039769875, + "loss": 0.3054, + "step": 35050 + }, + { + "epoch": 1.3537202208579482, + "grad_norm": 0.9203600883483887, + "learning_rate": 0.00010975455937809698, + "loss": 0.253, + "step": 35060 + }, + { + "epoch": 1.354106336151975, + "grad_norm": 0.33198195695877075, + "learning_rate": 0.00010972881835849518, + "loss": 0.3885, + "step": 35070 + }, + { + "epoch": 1.3544924514460017, + "grad_norm": 0.3201223611831665, + "learning_rate": 0.0001097030773388934, + "loss": 0.3029, + "step": 35080 + }, + { + "epoch": 1.3548785667400285, + "grad_norm": 1.2589943408966064, + "learning_rate": 0.0001096773363192916, + "loss": 0.4243, + "step": 35090 + }, + { + "epoch": 1.3552646820340555, + "grad_norm": 1.5106219053268433, + "learning_rate": 0.00010965159529968983, + "loss": 0.2585, + "step": 35100 + }, + { + "epoch": 1.3556507973280822, + "grad_norm": 1.429799199104309, + "learning_rate": 0.00010962585428008804, + "loss": 0.1961, + "step": 35110 + }, + { + "epoch": 1.356036912622109, + "grad_norm": 2.1211297512054443, + "learning_rate": 0.00010960011326048624, + "loss": 0.4057, + "step": 35120 + }, + { + "epoch": 1.3564230279161358, + "grad_norm": 2.5154731273651123, + "learning_rate": 0.00010957437224088447, + "loss": 0.3787, + "step": 35130 + }, + { + "epoch": 1.3568091432101625, + "grad_norm": 0.4914834201335907, + "learning_rate": 0.00010954863122128267, + "loss": 0.234, + "step": 35140 + }, + { + "epoch": 1.3571952585041893, + "grad_norm": 0.26685893535614014, + "learning_rate": 0.0001095228902016809, + "loss": 0.2841, + "step": 35150 + }, + { + "epoch": 1.357581373798216, + "grad_norm": 0.15462155640125275, + "learning_rate": 0.00010949714918207912, + "loss": 0.2269, + "step": 35160 + }, + { + "epoch": 1.357967489092243, + "grad_norm": 1.3887063264846802, + "learning_rate": 0.00010947140816247732, + "loss": 0.3455, + "step": 35170 + }, + { + "epoch": 1.3583536043862696, + "grad_norm": 0.786374032497406, + "learning_rate": 0.00010944566714287554, + "loss": 0.2897, + "step": 35180 + }, + { + "epoch": 1.3587397196802966, + "grad_norm": 1.100475549697876, + "learning_rate": 0.00010941992612327374, + "loss": 0.2892, + "step": 35190 + }, + { + "epoch": 1.3591258349743234, + "grad_norm": 0.7676102519035339, + "learning_rate": 0.00010939418510367196, + "loss": 0.1942, + "step": 35200 + }, + { + "epoch": 1.3595119502683501, + "grad_norm": 0.33462053537368774, + "learning_rate": 0.00010936844408407016, + "loss": 0.2872, + "step": 35210 + }, + { + "epoch": 1.359898065562377, + "grad_norm": 0.9294387698173523, + "learning_rate": 0.00010934270306446839, + "loss": 0.2617, + "step": 35220 + }, + { + "epoch": 1.3602841808564037, + "grad_norm": 0.3169979453086853, + "learning_rate": 0.00010931696204486662, + "loss": 0.2942, + "step": 35230 + }, + { + "epoch": 1.3606702961504304, + "grad_norm": 2.1339616775512695, + "learning_rate": 0.00010929122102526482, + "loss": 0.4448, + "step": 35240 + }, + { + "epoch": 1.3610564114444572, + "grad_norm": 0.9430062770843506, + "learning_rate": 0.00010926548000566303, + "loss": 0.2051, + "step": 35250 + }, + { + "epoch": 1.3614425267384842, + "grad_norm": 3.1187360286712646, + "learning_rate": 0.00010923973898606123, + "loss": 0.2274, + "step": 35260 + }, + { + "epoch": 1.361828642032511, + "grad_norm": 1.4727579355239868, + "learning_rate": 0.00010921399796645946, + "loss": 0.3757, + "step": 35270 + }, + { + "epoch": 1.3622147573265377, + "grad_norm": 2.157560348510742, + "learning_rate": 0.00010918825694685768, + "loss": 0.3096, + "step": 35280 + }, + { + "epoch": 1.3626008726205645, + "grad_norm": 0.33457377552986145, + "learning_rate": 0.00010916251592725588, + "loss": 0.1489, + "step": 35290 + }, + { + "epoch": 1.3629869879145913, + "grad_norm": 0.9005904197692871, + "learning_rate": 0.00010913677490765411, + "loss": 0.1826, + "step": 35300 + }, + { + "epoch": 1.363373103208618, + "grad_norm": 2.1222829818725586, + "learning_rate": 0.00010911103388805231, + "loss": 0.1965, + "step": 35310 + }, + { + "epoch": 1.3637592185026448, + "grad_norm": 1.3881357908248901, + "learning_rate": 0.00010908529286845052, + "loss": 0.1791, + "step": 35320 + }, + { + "epoch": 1.3641453337966718, + "grad_norm": 1.7574503421783447, + "learning_rate": 0.00010905955184884872, + "loss": 0.3316, + "step": 35330 + }, + { + "epoch": 1.3645314490906983, + "grad_norm": 0.1967727392911911, + "learning_rate": 0.00010903381082924695, + "loss": 0.2331, + "step": 35340 + }, + { + "epoch": 1.3649175643847253, + "grad_norm": 0.8974360823631287, + "learning_rate": 0.00010900806980964518, + "loss": 0.2589, + "step": 35350 + }, + { + "epoch": 1.365303679678752, + "grad_norm": 2.0996744632720947, + "learning_rate": 0.00010898232879004338, + "loss": 0.3663, + "step": 35360 + }, + { + "epoch": 1.3656897949727789, + "grad_norm": 0.5678316354751587, + "learning_rate": 0.0001089565877704416, + "loss": 0.1729, + "step": 35370 + }, + { + "epoch": 1.3660759102668056, + "grad_norm": 2.3381874561309814, + "learning_rate": 0.0001089308467508398, + "loss": 0.1615, + "step": 35380 + }, + { + "epoch": 1.3664620255608324, + "grad_norm": 1.0276836156845093, + "learning_rate": 0.00010890510573123802, + "loss": 0.3359, + "step": 35390 + }, + { + "epoch": 1.3668481408548594, + "grad_norm": 2.4374940395355225, + "learning_rate": 0.00010887936471163622, + "loss": 0.2435, + "step": 35400 + }, + { + "epoch": 1.367234256148886, + "grad_norm": 0.45221665501594543, + "learning_rate": 0.00010885362369203444, + "loss": 0.2555, + "step": 35410 + }, + { + "epoch": 1.367620371442913, + "grad_norm": 2.608090400695801, + "learning_rate": 0.00010882788267243267, + "loss": 0.2465, + "step": 35420 + }, + { + "epoch": 1.3680064867369397, + "grad_norm": 1.4186642169952393, + "learning_rate": 0.00010880214165283087, + "loss": 0.1674, + "step": 35430 + }, + { + "epoch": 1.3683926020309665, + "grad_norm": 0.659479022026062, + "learning_rate": 0.0001087764006332291, + "loss": 0.2926, + "step": 35440 + }, + { + "epoch": 1.3687787173249932, + "grad_norm": 0.9219567179679871, + "learning_rate": 0.0001087506596136273, + "loss": 0.2001, + "step": 35450 + }, + { + "epoch": 1.36916483261902, + "grad_norm": 0.8070804476737976, + "learning_rate": 0.00010872491859402551, + "loss": 0.2178, + "step": 35460 + }, + { + "epoch": 1.3695509479130468, + "grad_norm": 2.9981069564819336, + "learning_rate": 0.00010869917757442374, + "loss": 0.3079, + "step": 35470 + }, + { + "epoch": 1.3699370632070735, + "grad_norm": 0.7891242504119873, + "learning_rate": 0.00010867343655482194, + "loss": 0.2765, + "step": 35480 + }, + { + "epoch": 1.3703231785011005, + "grad_norm": 1.448637843132019, + "learning_rate": 0.00010864769553522016, + "loss": 0.3521, + "step": 35490 + }, + { + "epoch": 1.3707092937951273, + "grad_norm": 0.07628043740987778, + "learning_rate": 0.00010862195451561836, + "loss": 0.2083, + "step": 35500 + }, + { + "epoch": 1.371095409089154, + "grad_norm": 0.7549735307693481, + "learning_rate": 0.00010859621349601659, + "loss": 0.2536, + "step": 35510 + }, + { + "epoch": 1.3714815243831808, + "grad_norm": 1.3548041582107544, + "learning_rate": 0.00010857047247641479, + "loss": 0.251, + "step": 35520 + }, + { + "epoch": 1.3718676396772076, + "grad_norm": 0.530010998249054, + "learning_rate": 0.000108544731456813, + "loss": 0.1917, + "step": 35530 + }, + { + "epoch": 1.3722537549712344, + "grad_norm": 0.4148992896080017, + "learning_rate": 0.00010851899043721123, + "loss": 0.335, + "step": 35540 + }, + { + "epoch": 1.3726398702652611, + "grad_norm": 1.5118776559829712, + "learning_rate": 0.00010849324941760943, + "loss": 0.2159, + "step": 35550 + }, + { + "epoch": 1.3730259855592881, + "grad_norm": 1.036889910697937, + "learning_rate": 0.00010846750839800766, + "loss": 0.2975, + "step": 35560 + }, + { + "epoch": 1.3734121008533147, + "grad_norm": 1.724263072013855, + "learning_rate": 0.00010844176737840586, + "loss": 0.1476, + "step": 35570 + }, + { + "epoch": 1.3737982161473417, + "grad_norm": 1.599007487297058, + "learning_rate": 0.00010841602635880408, + "loss": 0.2539, + "step": 35580 + }, + { + "epoch": 1.3741843314413684, + "grad_norm": 2.9119279384613037, + "learning_rate": 0.00010839028533920228, + "loss": 0.2688, + "step": 35590 + }, + { + "epoch": 1.3745704467353952, + "grad_norm": 1.8647874593734741, + "learning_rate": 0.00010836454431960051, + "loss": 0.4158, + "step": 35600 + }, + { + "epoch": 1.374956562029422, + "grad_norm": 3.925290822982788, + "learning_rate": 0.00010833880329999872, + "loss": 0.3333, + "step": 35610 + }, + { + "epoch": 1.3753426773234487, + "grad_norm": 0.7124634385108948, + "learning_rate": 0.00010831306228039692, + "loss": 0.1069, + "step": 35620 + }, + { + "epoch": 1.3757287926174757, + "grad_norm": 1.303579330444336, + "learning_rate": 0.00010828732126079515, + "loss": 0.2898, + "step": 35630 + }, + { + "epoch": 1.3761149079115023, + "grad_norm": 3.921804189682007, + "learning_rate": 0.00010826158024119335, + "loss": 0.4212, + "step": 35640 + }, + { + "epoch": 1.3765010232055293, + "grad_norm": 1.3194564580917358, + "learning_rate": 0.00010823583922159158, + "loss": 0.2771, + "step": 35650 + }, + { + "epoch": 1.376887138499556, + "grad_norm": 1.4237637519836426, + "learning_rate": 0.00010821009820198979, + "loss": 0.2463, + "step": 35660 + }, + { + "epoch": 1.3772732537935828, + "grad_norm": 1.8165888786315918, + "learning_rate": 0.000108184357182388, + "loss": 0.291, + "step": 35670 + }, + { + "epoch": 1.3776593690876096, + "grad_norm": 1.1056426763534546, + "learning_rate": 0.00010815861616278622, + "loss": 0.2525, + "step": 35680 + }, + { + "epoch": 1.3780454843816363, + "grad_norm": 1.483189582824707, + "learning_rate": 0.00010813287514318442, + "loss": 0.1569, + "step": 35690 + }, + { + "epoch": 1.378431599675663, + "grad_norm": 1.0666841268539429, + "learning_rate": 0.00010810713412358264, + "loss": 0.235, + "step": 35700 + }, + { + "epoch": 1.3788177149696899, + "grad_norm": 1.0299845933914185, + "learning_rate": 0.00010808139310398084, + "loss": 0.3892, + "step": 35710 + }, + { + "epoch": 1.3792038302637168, + "grad_norm": 2.3474409580230713, + "learning_rate": 0.00010805565208437907, + "loss": 0.3417, + "step": 35720 + }, + { + "epoch": 1.3795899455577436, + "grad_norm": 1.7456315755844116, + "learning_rate": 0.0001080299110647773, + "loss": 0.2538, + "step": 35730 + }, + { + "epoch": 1.3799760608517704, + "grad_norm": 2.866103410720825, + "learning_rate": 0.0001080041700451755, + "loss": 0.1619, + "step": 35740 + }, + { + "epoch": 1.3803621761457971, + "grad_norm": 0.29136407375335693, + "learning_rate": 0.00010797842902557371, + "loss": 0.2692, + "step": 35750 + }, + { + "epoch": 1.380748291439824, + "grad_norm": 0.8046161532402039, + "learning_rate": 0.00010795268800597191, + "loss": 0.1575, + "step": 35760 + }, + { + "epoch": 1.3811344067338507, + "grad_norm": 0.6451787352561951, + "learning_rate": 0.00010792694698637014, + "loss": 0.4914, + "step": 35770 + }, + { + "epoch": 1.3815205220278775, + "grad_norm": 0.7289161086082458, + "learning_rate": 0.00010790120596676836, + "loss": 0.1895, + "step": 35780 + }, + { + "epoch": 1.3819066373219044, + "grad_norm": 0.8300430178642273, + "learning_rate": 0.00010787546494716656, + "loss": 0.3663, + "step": 35790 + }, + { + "epoch": 1.382292752615931, + "grad_norm": 0.17713364958763123, + "learning_rate": 0.00010784972392756479, + "loss": 0.3189, + "step": 35800 + }, + { + "epoch": 1.382678867909958, + "grad_norm": 0.903222918510437, + "learning_rate": 0.00010782398290796299, + "loss": 0.1577, + "step": 35810 + }, + { + "epoch": 1.3830649832039847, + "grad_norm": 0.08617932349443436, + "learning_rate": 0.0001077982418883612, + "loss": 0.2872, + "step": 35820 + }, + { + "epoch": 1.3834510984980115, + "grad_norm": 1.9590895175933838, + "learning_rate": 0.0001077725008687594, + "loss": 0.2907, + "step": 35830 + }, + { + "epoch": 1.3838372137920383, + "grad_norm": 1.2515161037445068, + "learning_rate": 0.00010774675984915763, + "loss": 0.177, + "step": 35840 + }, + { + "epoch": 1.384223329086065, + "grad_norm": 1.6171292066574097, + "learning_rate": 0.00010772101882955586, + "loss": 0.2321, + "step": 35850 + }, + { + "epoch": 1.3846094443800918, + "grad_norm": 0.13681405782699585, + "learning_rate": 0.00010769527780995406, + "loss": 0.247, + "step": 35860 + }, + { + "epoch": 1.3849955596741186, + "grad_norm": 1.1949968338012695, + "learning_rate": 0.00010766953679035228, + "loss": 0.2978, + "step": 35870 + }, + { + "epoch": 1.3853816749681456, + "grad_norm": 0.17001692950725555, + "learning_rate": 0.00010764379577075048, + "loss": 0.3241, + "step": 35880 + }, + { + "epoch": 1.3857677902621723, + "grad_norm": 0.8227952122688293, + "learning_rate": 0.0001076180547511487, + "loss": 0.3499, + "step": 35890 + }, + { + "epoch": 1.386153905556199, + "grad_norm": 1.4185482263565063, + "learning_rate": 0.0001075923137315469, + "loss": 0.3109, + "step": 35900 + }, + { + "epoch": 1.3865400208502259, + "grad_norm": 0.9533351063728333, + "learning_rate": 0.00010756657271194512, + "loss": 0.24, + "step": 35910 + }, + { + "epoch": 1.3869261361442526, + "grad_norm": 1.077789306640625, + "learning_rate": 0.00010754083169234335, + "loss": 0.2662, + "step": 35920 + }, + { + "epoch": 1.3873122514382794, + "grad_norm": 1.3528363704681396, + "learning_rate": 0.00010751509067274155, + "loss": 0.1623, + "step": 35930 + }, + { + "epoch": 1.3876983667323062, + "grad_norm": 0.25122806429862976, + "learning_rate": 0.00010748934965313978, + "loss": 0.1639, + "step": 35940 + }, + { + "epoch": 1.3880844820263332, + "grad_norm": 0.9446159601211548, + "learning_rate": 0.00010746360863353798, + "loss": 0.2035, + "step": 35950 + }, + { + "epoch": 1.38847059732036, + "grad_norm": 1.2258719205856323, + "learning_rate": 0.00010743786761393619, + "loss": 0.4247, + "step": 35960 + }, + { + "epoch": 1.3888567126143867, + "grad_norm": 3.117729663848877, + "learning_rate": 0.00010741212659433442, + "loss": 0.2569, + "step": 35970 + }, + { + "epoch": 1.3892428279084135, + "grad_norm": 0.6333123445510864, + "learning_rate": 0.00010738638557473262, + "loss": 0.1843, + "step": 35980 + }, + { + "epoch": 1.3896289432024402, + "grad_norm": 1.49360191822052, + "learning_rate": 0.00010736064455513084, + "loss": 0.2318, + "step": 35990 + }, + { + "epoch": 1.390015058496467, + "grad_norm": 3.9082753658294678, + "learning_rate": 0.00010733490353552904, + "loss": 0.2037, + "step": 36000 + }, + { + "epoch": 1.3904011737904938, + "grad_norm": 0.5687323808670044, + "learning_rate": 0.00010730916251592727, + "loss": 0.0871, + "step": 36010 + }, + { + "epoch": 1.3907872890845208, + "grad_norm": 0.3480868339538574, + "learning_rate": 0.00010728342149632547, + "loss": 0.4009, + "step": 36020 + }, + { + "epoch": 1.3911734043785473, + "grad_norm": 1.325042486190796, + "learning_rate": 0.00010725768047672368, + "loss": 0.5143, + "step": 36030 + }, + { + "epoch": 1.3915595196725743, + "grad_norm": 2.114786386489868, + "learning_rate": 0.00010723193945712191, + "loss": 0.2944, + "step": 36040 + }, + { + "epoch": 1.391945634966601, + "grad_norm": 1.716272234916687, + "learning_rate": 0.00010720619843752011, + "loss": 0.3793, + "step": 36050 + }, + { + "epoch": 1.3923317502606278, + "grad_norm": 2.057535171508789, + "learning_rate": 0.00010718045741791834, + "loss": 0.199, + "step": 36060 + }, + { + "epoch": 1.3927178655546546, + "grad_norm": 2.371248483657837, + "learning_rate": 0.00010715471639831653, + "loss": 0.3276, + "step": 36070 + }, + { + "epoch": 1.3931039808486814, + "grad_norm": 0.1784186065196991, + "learning_rate": 0.00010712897537871476, + "loss": 0.0922, + "step": 36080 + }, + { + "epoch": 1.3934900961427081, + "grad_norm": 0.7923040390014648, + "learning_rate": 0.00010710323435911298, + "loss": 0.1599, + "step": 36090 + }, + { + "epoch": 1.393876211436735, + "grad_norm": 1.9013831615447998, + "learning_rate": 0.00010707749333951117, + "loss": 0.4997, + "step": 36100 + }, + { + "epoch": 1.394262326730762, + "grad_norm": 1.8659415245056152, + "learning_rate": 0.0001070517523199094, + "loss": 0.1851, + "step": 36110 + }, + { + "epoch": 1.3946484420247887, + "grad_norm": 2.5775375366210938, + "learning_rate": 0.0001070260113003076, + "loss": 0.2384, + "step": 36120 + }, + { + "epoch": 1.3950345573188154, + "grad_norm": 0.21943879127502441, + "learning_rate": 0.00010700027028070583, + "loss": 0.3996, + "step": 36130 + }, + { + "epoch": 1.3954206726128422, + "grad_norm": 1.1734743118286133, + "learning_rate": 0.00010697452926110403, + "loss": 0.1737, + "step": 36140 + }, + { + "epoch": 1.395806787906869, + "grad_norm": 0.697695791721344, + "learning_rate": 0.00010694878824150225, + "loss": 0.2778, + "step": 36150 + }, + { + "epoch": 1.3961929032008957, + "grad_norm": 3.2881579399108887, + "learning_rate": 0.00010692304722190047, + "loss": 0.3198, + "step": 36160 + }, + { + "epoch": 1.3965790184949225, + "grad_norm": 0.1592467725276947, + "learning_rate": 0.00010689730620229868, + "loss": 0.2591, + "step": 36170 + }, + { + "epoch": 1.3969651337889495, + "grad_norm": 0.4579029083251953, + "learning_rate": 0.0001068715651826969, + "loss": 0.2984, + "step": 36180 + }, + { + "epoch": 1.3973512490829763, + "grad_norm": 0.4286015033721924, + "learning_rate": 0.0001068458241630951, + "loss": 0.4098, + "step": 36190 + }, + { + "epoch": 1.397737364377003, + "grad_norm": 1.7824127674102783, + "learning_rate": 0.00010682008314349332, + "loss": 0.2446, + "step": 36200 + }, + { + "epoch": 1.3981234796710298, + "grad_norm": 0.8584449887275696, + "learning_rate": 0.00010679434212389152, + "loss": 0.1415, + "step": 36210 + }, + { + "epoch": 1.3985095949650566, + "grad_norm": 1.1699339151382446, + "learning_rate": 0.00010676860110428975, + "loss": 0.1706, + "step": 36220 + }, + { + "epoch": 1.3988957102590833, + "grad_norm": 2.615877389907837, + "learning_rate": 0.00010674286008468796, + "loss": 0.2605, + "step": 36230 + }, + { + "epoch": 1.39928182555311, + "grad_norm": 2.182037591934204, + "learning_rate": 0.00010671711906508617, + "loss": 0.2067, + "step": 36240 + }, + { + "epoch": 1.399667940847137, + "grad_norm": 2.183263063430786, + "learning_rate": 0.00010669137804548439, + "loss": 0.4704, + "step": 36250 + }, + { + "epoch": 1.4000540561411636, + "grad_norm": 3.505791187286377, + "learning_rate": 0.00010666563702588259, + "loss": 0.3385, + "step": 36260 + }, + { + "epoch": 1.4004401714351906, + "grad_norm": 1.2262030839920044, + "learning_rate": 0.00010663989600628081, + "loss": 0.3999, + "step": 36270 + }, + { + "epoch": 1.4008262867292174, + "grad_norm": 2.4024577140808105, + "learning_rate": 0.00010661415498667904, + "loss": 0.1984, + "step": 36280 + }, + { + "epoch": 1.4012124020232442, + "grad_norm": 0.4166090786457062, + "learning_rate": 0.00010658841396707724, + "loss": 0.2448, + "step": 36290 + }, + { + "epoch": 1.401598517317271, + "grad_norm": 0.422590047121048, + "learning_rate": 0.00010656267294747547, + "loss": 0.2464, + "step": 36300 + }, + { + "epoch": 1.4019846326112977, + "grad_norm": 2.287503480911255, + "learning_rate": 0.00010653693192787367, + "loss": 0.1621, + "step": 36310 + }, + { + "epoch": 1.4023707479053245, + "grad_norm": 0.8126110434532166, + "learning_rate": 0.00010651119090827188, + "loss": 0.2097, + "step": 36320 + }, + { + "epoch": 1.4027568631993512, + "grad_norm": 0.683016836643219, + "learning_rate": 0.00010648544988867008, + "loss": 0.1512, + "step": 36330 + }, + { + "epoch": 1.4031429784933782, + "grad_norm": 1.5477893352508545, + "learning_rate": 0.00010645970886906831, + "loss": 0.2169, + "step": 36340 + }, + { + "epoch": 1.403529093787405, + "grad_norm": 2.183166265487671, + "learning_rate": 0.00010643396784946653, + "loss": 0.4307, + "step": 36350 + }, + { + "epoch": 1.4039152090814317, + "grad_norm": 1.5782747268676758, + "learning_rate": 0.00010640822682986473, + "loss": 0.291, + "step": 36360 + }, + { + "epoch": 1.4043013243754585, + "grad_norm": 1.2823392152786255, + "learning_rate": 0.00010638248581026296, + "loss": 0.315, + "step": 36370 + }, + { + "epoch": 1.4046874396694853, + "grad_norm": 2.0126500129699707, + "learning_rate": 0.00010635674479066116, + "loss": 0.2565, + "step": 36380 + }, + { + "epoch": 1.405073554963512, + "grad_norm": 2.490217447280884, + "learning_rate": 0.00010633100377105937, + "loss": 0.2665, + "step": 36390 + }, + { + "epoch": 1.4054596702575388, + "grad_norm": 1.2054855823516846, + "learning_rate": 0.00010630526275145757, + "loss": 0.5455, + "step": 36400 + }, + { + "epoch": 1.4058457855515658, + "grad_norm": 0.7968757748603821, + "learning_rate": 0.0001062795217318558, + "loss": 0.333, + "step": 36410 + }, + { + "epoch": 1.4062319008455926, + "grad_norm": 1.1027718782424927, + "learning_rate": 0.00010625378071225403, + "loss": 0.2019, + "step": 36420 + }, + { + "epoch": 1.4066180161396193, + "grad_norm": 1.987302541732788, + "learning_rate": 0.00010622803969265223, + "loss": 0.3159, + "step": 36430 + }, + { + "epoch": 1.407004131433646, + "grad_norm": 1.5426512956619263, + "learning_rate": 0.00010620229867305045, + "loss": 0.3759, + "step": 36440 + }, + { + "epoch": 1.4073902467276729, + "grad_norm": 0.5228156447410583, + "learning_rate": 0.00010617655765344865, + "loss": 0.2465, + "step": 36450 + }, + { + "epoch": 1.4077763620216996, + "grad_norm": 0.46890121698379517, + "learning_rate": 0.00010615081663384687, + "loss": 0.302, + "step": 36460 + }, + { + "epoch": 1.4081624773157264, + "grad_norm": 2.1506495475769043, + "learning_rate": 0.0001061250756142451, + "loss": 0.2569, + "step": 36470 + }, + { + "epoch": 1.4085485926097534, + "grad_norm": 2.307468891143799, + "learning_rate": 0.0001060993345946433, + "loss": 0.2009, + "step": 36480 + }, + { + "epoch": 1.40893470790378, + "grad_norm": 0.07033026963472366, + "learning_rate": 0.00010607359357504152, + "loss": 0.1728, + "step": 36490 + }, + { + "epoch": 1.409320823197807, + "grad_norm": 0.3262972831726074, + "learning_rate": 0.00010604785255543972, + "loss": 0.2905, + "step": 36500 + }, + { + "epoch": 1.4097069384918337, + "grad_norm": 0.755646824836731, + "learning_rate": 0.00010602211153583795, + "loss": 0.3287, + "step": 36510 + }, + { + "epoch": 1.4100930537858605, + "grad_norm": 1.1180161237716675, + "learning_rate": 0.00010599637051623615, + "loss": 0.2743, + "step": 36520 + }, + { + "epoch": 1.4104791690798872, + "grad_norm": 1.2358392477035522, + "learning_rate": 0.00010597062949663436, + "loss": 0.2672, + "step": 36530 + }, + { + "epoch": 1.410865284373914, + "grad_norm": 1.4188988208770752, + "learning_rate": 0.00010594488847703259, + "loss": 0.2552, + "step": 36540 + }, + { + "epoch": 1.4112513996679408, + "grad_norm": 1.4727978706359863, + "learning_rate": 0.00010591914745743079, + "loss": 0.2256, + "step": 36550 + }, + { + "epoch": 1.4116375149619675, + "grad_norm": 0.08973213285207748, + "learning_rate": 0.00010589340643782901, + "loss": 0.224, + "step": 36560 + }, + { + "epoch": 1.4120236302559945, + "grad_norm": 0.9915102124214172, + "learning_rate": 0.00010586766541822721, + "loss": 0.195, + "step": 36570 + }, + { + "epoch": 1.4124097455500213, + "grad_norm": 0.8524800539016724, + "learning_rate": 0.00010584192439862544, + "loss": 0.1492, + "step": 36580 + }, + { + "epoch": 1.412795860844048, + "grad_norm": 1.8414466381072998, + "learning_rate": 0.00010581618337902365, + "loss": 0.3412, + "step": 36590 + }, + { + "epoch": 1.4131819761380748, + "grad_norm": 2.596547842025757, + "learning_rate": 0.00010579044235942185, + "loss": 0.3469, + "step": 36600 + }, + { + "epoch": 1.4135680914321016, + "grad_norm": 0.5367813110351562, + "learning_rate": 0.00010576470133982008, + "loss": 0.2274, + "step": 36610 + }, + { + "epoch": 1.4139542067261284, + "grad_norm": 2.70858097076416, + "learning_rate": 0.00010573896032021828, + "loss": 0.171, + "step": 36620 + }, + { + "epoch": 1.4143403220201551, + "grad_norm": 1.7077667713165283, + "learning_rate": 0.00010571321930061651, + "loss": 0.2638, + "step": 36630 + }, + { + "epoch": 1.4147264373141821, + "grad_norm": 0.9189953804016113, + "learning_rate": 0.00010568747828101471, + "loss": 0.2283, + "step": 36640 + }, + { + "epoch": 1.4151125526082087, + "grad_norm": 2.0240087509155273, + "learning_rate": 0.00010566173726141293, + "loss": 0.3166, + "step": 36650 + }, + { + "epoch": 1.4154986679022357, + "grad_norm": 3.3304507732391357, + "learning_rate": 0.00010563599624181115, + "loss": 0.3046, + "step": 36660 + }, + { + "epoch": 1.4158847831962624, + "grad_norm": 1.955929160118103, + "learning_rate": 0.00010561025522220935, + "loss": 0.2653, + "step": 36670 + }, + { + "epoch": 1.4162708984902892, + "grad_norm": 2.2465381622314453, + "learning_rate": 0.00010558451420260757, + "loss": 0.3119, + "step": 36680 + }, + { + "epoch": 1.416657013784316, + "grad_norm": 2.108614921569824, + "learning_rate": 0.00010555877318300577, + "loss": 0.2872, + "step": 36690 + }, + { + "epoch": 1.4170431290783427, + "grad_norm": 1.35493004322052, + "learning_rate": 0.000105533032163404, + "loss": 0.1491, + "step": 36700 + }, + { + "epoch": 1.4174292443723697, + "grad_norm": 0.9102393984794617, + "learning_rate": 0.0001055072911438022, + "loss": 0.1565, + "step": 36710 + }, + { + "epoch": 1.4178153596663963, + "grad_norm": 1.7878345251083374, + "learning_rate": 0.00010548155012420043, + "loss": 0.4864, + "step": 36720 + }, + { + "epoch": 1.4182014749604233, + "grad_norm": 1.4333637952804565, + "learning_rate": 0.00010545580910459864, + "loss": 0.3246, + "step": 36730 + }, + { + "epoch": 1.41858759025445, + "grad_norm": 0.0966360941529274, + "learning_rate": 0.00010543006808499684, + "loss": 0.1758, + "step": 36740 + }, + { + "epoch": 1.4189737055484768, + "grad_norm": 0.30043545365333557, + "learning_rate": 0.00010540432706539507, + "loss": 0.1811, + "step": 36750 + }, + { + "epoch": 1.4193598208425036, + "grad_norm": 1.7705951929092407, + "learning_rate": 0.00010537858604579327, + "loss": 0.1824, + "step": 36760 + }, + { + "epoch": 1.4197459361365303, + "grad_norm": 1.2025195360183716, + "learning_rate": 0.0001053528450261915, + "loss": 0.1463, + "step": 36770 + }, + { + "epoch": 1.420132051430557, + "grad_norm": 0.3154304325580597, + "learning_rate": 0.00010532710400658972, + "loss": 0.1817, + "step": 36780 + }, + { + "epoch": 1.4205181667245839, + "grad_norm": 3.392331838607788, + "learning_rate": 0.00010530136298698792, + "loss": 0.2938, + "step": 36790 + }, + { + "epoch": 1.4209042820186109, + "grad_norm": 0.7256132364273071, + "learning_rate": 0.00010527562196738613, + "loss": 0.28, + "step": 36800 + }, + { + "epoch": 1.4212903973126376, + "grad_norm": 1.9007991552352905, + "learning_rate": 0.00010524988094778435, + "loss": 0.2306, + "step": 36810 + }, + { + "epoch": 1.4216765126066644, + "grad_norm": 3.21189546585083, + "learning_rate": 0.00010522413992818256, + "loss": 0.2945, + "step": 36820 + }, + { + "epoch": 1.4220626279006912, + "grad_norm": 0.20476600527763367, + "learning_rate": 0.00010519839890858076, + "loss": 0.3129, + "step": 36830 + }, + { + "epoch": 1.422448743194718, + "grad_norm": 1.4225107431411743, + "learning_rate": 0.00010517265788897899, + "loss": 0.2262, + "step": 36840 + }, + { + "epoch": 1.4228348584887447, + "grad_norm": 1.203728437423706, + "learning_rate": 0.00010514691686937721, + "loss": 0.176, + "step": 36850 + }, + { + "epoch": 1.4232209737827715, + "grad_norm": 0.3001759648323059, + "learning_rate": 0.00010512117584977541, + "loss": 0.218, + "step": 36860 + }, + { + "epoch": 1.4236070890767984, + "grad_norm": 2.3143389225006104, + "learning_rate": 0.00010509543483017363, + "loss": 0.1562, + "step": 36870 + }, + { + "epoch": 1.423993204370825, + "grad_norm": 0.5087364912033081, + "learning_rate": 0.00010506969381057184, + "loss": 0.1207, + "step": 36880 + }, + { + "epoch": 1.424379319664852, + "grad_norm": 1.6521960496902466, + "learning_rate": 0.00010504395279097005, + "loss": 0.4756, + "step": 36890 + }, + { + "epoch": 1.4247654349588788, + "grad_norm": 0.09236706793308258, + "learning_rate": 0.00010501821177136825, + "loss": 0.1755, + "step": 36900 + }, + { + "epoch": 1.4251515502529055, + "grad_norm": 0.2143094390630722, + "learning_rate": 0.00010499247075176648, + "loss": 0.2126, + "step": 36910 + }, + { + "epoch": 1.4255376655469323, + "grad_norm": 1.210170865058899, + "learning_rate": 0.00010496672973216471, + "loss": 0.1391, + "step": 36920 + }, + { + "epoch": 1.425923780840959, + "grad_norm": 0.6248244047164917, + "learning_rate": 0.00010494098871256291, + "loss": 0.2898, + "step": 36930 + }, + { + "epoch": 1.426309896134986, + "grad_norm": 2.1100337505340576, + "learning_rate": 0.00010491524769296113, + "loss": 0.1198, + "step": 36940 + }, + { + "epoch": 1.4266960114290126, + "grad_norm": 2.5673348903656006, + "learning_rate": 0.00010488950667335933, + "loss": 0.2284, + "step": 36950 + }, + { + "epoch": 1.4270821267230396, + "grad_norm": 1.868195652961731, + "learning_rate": 0.00010486376565375755, + "loss": 0.3738, + "step": 36960 + }, + { + "epoch": 1.4274682420170663, + "grad_norm": 1.0951671600341797, + "learning_rate": 0.00010483802463415577, + "loss": 0.1424, + "step": 36970 + }, + { + "epoch": 1.4278543573110931, + "grad_norm": 0.4791143536567688, + "learning_rate": 0.00010481228361455397, + "loss": 0.1445, + "step": 36980 + }, + { + "epoch": 1.4282404726051199, + "grad_norm": 3.2477540969848633, + "learning_rate": 0.0001047865425949522, + "loss": 0.2676, + "step": 36990 + }, + { + "epoch": 1.4286265878991466, + "grad_norm": 0.8082342147827148, + "learning_rate": 0.0001047608015753504, + "loss": 0.2295, + "step": 37000 + }, + { + "epoch": 1.4290127031931734, + "grad_norm": 2.3048954010009766, + "learning_rate": 0.00010473506055574863, + "loss": 0.1996, + "step": 37010 + }, + { + "epoch": 1.4293988184872002, + "grad_norm": 0.42648783326148987, + "learning_rate": 0.00010470931953614683, + "loss": 0.2549, + "step": 37020 + }, + { + "epoch": 1.4297849337812272, + "grad_norm": 1.8794362545013428, + "learning_rate": 0.00010468357851654504, + "loss": 0.1929, + "step": 37030 + }, + { + "epoch": 1.430171049075254, + "grad_norm": 0.9679039120674133, + "learning_rate": 0.00010465783749694327, + "loss": 0.1544, + "step": 37040 + }, + { + "epoch": 1.4305571643692807, + "grad_norm": 0.7789367437362671, + "learning_rate": 0.00010463209647734147, + "loss": 0.2251, + "step": 37050 + }, + { + "epoch": 1.4309432796633075, + "grad_norm": 1.1705437898635864, + "learning_rate": 0.0001046063554577397, + "loss": 0.1567, + "step": 37060 + }, + { + "epoch": 1.4313293949573342, + "grad_norm": 0.13087430596351624, + "learning_rate": 0.00010458061443813789, + "loss": 0.3237, + "step": 37070 + }, + { + "epoch": 1.431715510251361, + "grad_norm": 1.0658761262893677, + "learning_rate": 0.00010455487341853612, + "loss": 0.24, + "step": 37080 + }, + { + "epoch": 1.4321016255453878, + "grad_norm": 1.686922311782837, + "learning_rate": 0.00010452913239893433, + "loss": 0.1911, + "step": 37090 + }, + { + "epoch": 1.4324877408394148, + "grad_norm": 2.010221481323242, + "learning_rate": 0.00010450339137933253, + "loss": 0.4569, + "step": 37100 + }, + { + "epoch": 1.4328738561334413, + "grad_norm": 0.8007562160491943, + "learning_rate": 0.00010447765035973076, + "loss": 0.1198, + "step": 37110 + }, + { + "epoch": 1.4332599714274683, + "grad_norm": 0.5455211997032166, + "learning_rate": 0.00010445190934012896, + "loss": 0.2179, + "step": 37120 + }, + { + "epoch": 1.433646086721495, + "grad_norm": 1.4133542776107788, + "learning_rate": 0.00010442616832052719, + "loss": 0.45, + "step": 37130 + }, + { + "epoch": 1.4340322020155218, + "grad_norm": 1.6218222379684448, + "learning_rate": 0.00010440042730092539, + "loss": 0.19, + "step": 37140 + }, + { + "epoch": 1.4344183173095486, + "grad_norm": 0.6425970196723938, + "learning_rate": 0.00010437468628132361, + "loss": 0.2345, + "step": 37150 + }, + { + "epoch": 1.4348044326035754, + "grad_norm": 0.4344918131828308, + "learning_rate": 0.00010434894526172183, + "loss": 0.2532, + "step": 37160 + }, + { + "epoch": 1.4351905478976021, + "grad_norm": 0.6628998517990112, + "learning_rate": 0.00010432320424212003, + "loss": 0.2281, + "step": 37170 + }, + { + "epoch": 1.435576663191629, + "grad_norm": 0.8695842027664185, + "learning_rate": 0.00010429746322251825, + "loss": 0.2796, + "step": 37180 + }, + { + "epoch": 1.435962778485656, + "grad_norm": 0.16481854021549225, + "learning_rate": 0.00010427172220291645, + "loss": 0.2333, + "step": 37190 + }, + { + "epoch": 1.4363488937796827, + "grad_norm": 0.7194454073905945, + "learning_rate": 0.00010424598118331468, + "loss": 0.1413, + "step": 37200 + }, + { + "epoch": 1.4367350090737094, + "grad_norm": 4.845508575439453, + "learning_rate": 0.00010422024016371288, + "loss": 0.3944, + "step": 37210 + }, + { + "epoch": 1.4371211243677362, + "grad_norm": 2.6386618614196777, + "learning_rate": 0.0001041944991441111, + "loss": 0.3037, + "step": 37220 + }, + { + "epoch": 1.437507239661763, + "grad_norm": 0.4089922308921814, + "learning_rate": 0.00010416875812450932, + "loss": 0.2985, + "step": 37230 + }, + { + "epoch": 1.4378933549557897, + "grad_norm": 1.456944465637207, + "learning_rate": 0.00010414301710490752, + "loss": 0.3031, + "step": 37240 + }, + { + "epoch": 1.4382794702498165, + "grad_norm": 1.301829218864441, + "learning_rate": 0.00010411727608530575, + "loss": 0.2578, + "step": 37250 + }, + { + "epoch": 1.4386655855438435, + "grad_norm": 1.2072703838348389, + "learning_rate": 0.00010409153506570395, + "loss": 0.391, + "step": 37260 + }, + { + "epoch": 1.4390517008378703, + "grad_norm": 0.5538531541824341, + "learning_rate": 0.00010406579404610217, + "loss": 0.1867, + "step": 37270 + }, + { + "epoch": 1.439437816131897, + "grad_norm": 1.0898678302764893, + "learning_rate": 0.0001040400530265004, + "loss": 0.2112, + "step": 37280 + }, + { + "epoch": 1.4398239314259238, + "grad_norm": 1.5769239664077759, + "learning_rate": 0.0001040143120068986, + "loss": 0.3121, + "step": 37290 + }, + { + "epoch": 1.4402100467199506, + "grad_norm": 0.3964422345161438, + "learning_rate": 0.00010398857098729681, + "loss": 0.196, + "step": 37300 + }, + { + "epoch": 1.4405961620139773, + "grad_norm": 1.0268182754516602, + "learning_rate": 0.00010396282996769501, + "loss": 0.1785, + "step": 37310 + }, + { + "epoch": 1.440982277308004, + "grad_norm": 1.750826358795166, + "learning_rate": 0.00010393708894809324, + "loss": 0.3654, + "step": 37320 + }, + { + "epoch": 1.441368392602031, + "grad_norm": 1.1231745481491089, + "learning_rate": 0.00010391134792849144, + "loss": 0.2594, + "step": 37330 + }, + { + "epoch": 1.4417545078960576, + "grad_norm": 0.2897786498069763, + "learning_rate": 0.00010388560690888967, + "loss": 0.3483, + "step": 37340 + }, + { + "epoch": 1.4421406231900846, + "grad_norm": 0.07170752435922623, + "learning_rate": 0.00010385986588928789, + "loss": 0.1462, + "step": 37350 + }, + { + "epoch": 1.4425267384841114, + "grad_norm": 3.326099395751953, + "learning_rate": 0.00010383412486968609, + "loss": 0.3785, + "step": 37360 + }, + { + "epoch": 1.4429128537781382, + "grad_norm": 0.5063263773918152, + "learning_rate": 0.0001038083838500843, + "loss": 0.2974, + "step": 37370 + }, + { + "epoch": 1.443298969072165, + "grad_norm": 1.160088062286377, + "learning_rate": 0.00010378264283048252, + "loss": 0.3003, + "step": 37380 + }, + { + "epoch": 1.4436850843661917, + "grad_norm": 0.7043284177780151, + "learning_rate": 0.00010375690181088073, + "loss": 0.3193, + "step": 37390 + }, + { + "epoch": 1.4440711996602185, + "grad_norm": 0.2916620373725891, + "learning_rate": 0.00010373116079127896, + "loss": 0.235, + "step": 37400 + }, + { + "epoch": 1.4444573149542452, + "grad_norm": 2.1940219402313232, + "learning_rate": 0.00010370541977167716, + "loss": 0.2657, + "step": 37410 + }, + { + "epoch": 1.4448434302482722, + "grad_norm": 1.849794626235962, + "learning_rate": 0.00010367967875207539, + "loss": 0.3121, + "step": 37420 + }, + { + "epoch": 1.445229545542299, + "grad_norm": 1.1139589548110962, + "learning_rate": 0.00010365393773247359, + "loss": 0.2179, + "step": 37430 + }, + { + "epoch": 1.4456156608363258, + "grad_norm": 0.917142927646637, + "learning_rate": 0.0001036281967128718, + "loss": 0.2406, + "step": 37440 + }, + { + "epoch": 1.4460017761303525, + "grad_norm": 1.6973673105239868, + "learning_rate": 0.00010360245569327001, + "loss": 0.3236, + "step": 37450 + }, + { + "epoch": 1.4463878914243793, + "grad_norm": 0.3979933559894562, + "learning_rate": 0.00010357671467366823, + "loss": 0.1477, + "step": 37460 + }, + { + "epoch": 1.446774006718406, + "grad_norm": 0.8938451409339905, + "learning_rate": 0.00010355097365406645, + "loss": 0.116, + "step": 37470 + }, + { + "epoch": 1.4471601220124328, + "grad_norm": 1.068787693977356, + "learning_rate": 0.00010352523263446465, + "loss": 0.2342, + "step": 37480 + }, + { + "epoch": 1.4475462373064598, + "grad_norm": 1.1985591650009155, + "learning_rate": 0.00010349949161486288, + "loss": 0.2423, + "step": 37490 + }, + { + "epoch": 1.4479323526004866, + "grad_norm": 1.0383973121643066, + "learning_rate": 0.00010347375059526108, + "loss": 0.3059, + "step": 37500 + }, + { + "epoch": 1.4483184678945134, + "grad_norm": 0.26652297377586365, + "learning_rate": 0.0001034480095756593, + "loss": 0.262, + "step": 37510 + }, + { + "epoch": 1.4487045831885401, + "grad_norm": 1.8498083353042603, + "learning_rate": 0.0001034222685560575, + "loss": 0.2122, + "step": 37520 + }, + { + "epoch": 1.4490906984825669, + "grad_norm": 1.2896068096160889, + "learning_rate": 0.00010339652753645572, + "loss": 0.2939, + "step": 37530 + }, + { + "epoch": 1.4494768137765937, + "grad_norm": 1.423343300819397, + "learning_rate": 0.00010337078651685395, + "loss": 0.3375, + "step": 37540 + }, + { + "epoch": 1.4498629290706204, + "grad_norm": 1.8248246908187866, + "learning_rate": 0.00010334504549725215, + "loss": 0.3534, + "step": 37550 + }, + { + "epoch": 1.4502490443646474, + "grad_norm": 2.3713393211364746, + "learning_rate": 0.00010331930447765037, + "loss": 0.2795, + "step": 37560 + }, + { + "epoch": 1.450635159658674, + "grad_norm": 2.1431849002838135, + "learning_rate": 0.00010329356345804857, + "loss": 0.2503, + "step": 37570 + }, + { + "epoch": 1.451021274952701, + "grad_norm": 1.6521297693252563, + "learning_rate": 0.0001032678224384468, + "loss": 0.1789, + "step": 37580 + }, + { + "epoch": 1.4514073902467277, + "grad_norm": 1.5589754581451416, + "learning_rate": 0.00010324208141884501, + "loss": 0.3201, + "step": 37590 + }, + { + "epoch": 1.4517935055407545, + "grad_norm": 0.4339803159236908, + "learning_rate": 0.00010321634039924321, + "loss": 0.2964, + "step": 37600 + }, + { + "epoch": 1.4521796208347812, + "grad_norm": 0.30054792761802673, + "learning_rate": 0.00010319059937964144, + "loss": 0.3477, + "step": 37610 + }, + { + "epoch": 1.452565736128808, + "grad_norm": 2.2864038944244385, + "learning_rate": 0.00010316485836003964, + "loss": 0.2849, + "step": 37620 + }, + { + "epoch": 1.4529518514228348, + "grad_norm": 1.8392651081085205, + "learning_rate": 0.00010313911734043787, + "loss": 0.2089, + "step": 37630 + }, + { + "epoch": 1.4533379667168616, + "grad_norm": 1.0444347858428955, + "learning_rate": 0.00010311337632083607, + "loss": 0.2306, + "step": 37640 + }, + { + "epoch": 1.4537240820108885, + "grad_norm": 1.265647053718567, + "learning_rate": 0.00010308763530123429, + "loss": 0.2192, + "step": 37650 + }, + { + "epoch": 1.4541101973049153, + "grad_norm": 1.0464913845062256, + "learning_rate": 0.0001030618942816325, + "loss": 0.2857, + "step": 37660 + }, + { + "epoch": 1.454496312598942, + "grad_norm": 0.5931240320205688, + "learning_rate": 0.0001030361532620307, + "loss": 0.1751, + "step": 37670 + }, + { + "epoch": 1.4548824278929688, + "grad_norm": 2.0155346393585205, + "learning_rate": 0.00010301041224242893, + "loss": 0.4124, + "step": 37680 + }, + { + "epoch": 1.4552685431869956, + "grad_norm": 2.843345880508423, + "learning_rate": 0.00010298467122282713, + "loss": 0.2119, + "step": 37690 + }, + { + "epoch": 1.4556546584810224, + "grad_norm": 1.977612018585205, + "learning_rate": 0.00010295893020322536, + "loss": 0.271, + "step": 37700 + }, + { + "epoch": 1.4560407737750491, + "grad_norm": 1.2862237691879272, + "learning_rate": 0.00010293318918362356, + "loss": 0.4147, + "step": 37710 + }, + { + "epoch": 1.4564268890690761, + "grad_norm": 0.8235340118408203, + "learning_rate": 0.00010290744816402179, + "loss": 0.1918, + "step": 37720 + }, + { + "epoch": 1.456813004363103, + "grad_norm": 1.0351753234863281, + "learning_rate": 0.00010288170714442, + "loss": 0.2343, + "step": 37730 + }, + { + "epoch": 1.4571991196571297, + "grad_norm": 0.2567160725593567, + "learning_rate": 0.0001028559661248182, + "loss": 0.2664, + "step": 37740 + }, + { + "epoch": 1.4575852349511564, + "grad_norm": 2.7498669624328613, + "learning_rate": 0.00010283022510521643, + "loss": 0.2839, + "step": 37750 + }, + { + "epoch": 1.4579713502451832, + "grad_norm": 0.7125422358512878, + "learning_rate": 0.00010280448408561463, + "loss": 0.2411, + "step": 37760 + }, + { + "epoch": 1.45835746553921, + "grad_norm": 0.720761239528656, + "learning_rate": 0.00010277874306601285, + "loss": 0.315, + "step": 37770 + }, + { + "epoch": 1.4587435808332367, + "grad_norm": 3.188563346862793, + "learning_rate": 0.00010275300204641108, + "loss": 0.3945, + "step": 37780 + }, + { + "epoch": 1.4591296961272637, + "grad_norm": 1.23422110080719, + "learning_rate": 0.00010272726102680928, + "loss": 0.1828, + "step": 37790 + }, + { + "epoch": 1.4595158114212903, + "grad_norm": 1.1572456359863281, + "learning_rate": 0.00010270152000720749, + "loss": 0.1687, + "step": 37800 + }, + { + "epoch": 1.4599019267153173, + "grad_norm": 0.6565262079238892, + "learning_rate": 0.00010267577898760569, + "loss": 0.3145, + "step": 37810 + }, + { + "epoch": 1.460288042009344, + "grad_norm": 0.6239646673202515, + "learning_rate": 0.00010265003796800392, + "loss": 0.2381, + "step": 37820 + }, + { + "epoch": 1.4606741573033708, + "grad_norm": 0.7578912377357483, + "learning_rate": 0.00010262429694840212, + "loss": 0.2019, + "step": 37830 + }, + { + "epoch": 1.4610602725973976, + "grad_norm": 1.0978549718856812, + "learning_rate": 0.00010259855592880035, + "loss": 0.2477, + "step": 37840 + }, + { + "epoch": 1.4614463878914243, + "grad_norm": 0.3615519404411316, + "learning_rate": 0.00010257281490919857, + "loss": 0.4241, + "step": 37850 + }, + { + "epoch": 1.461832503185451, + "grad_norm": 0.10355047881603241, + "learning_rate": 0.00010254707388959677, + "loss": 0.1181, + "step": 37860 + }, + { + "epoch": 1.4622186184794779, + "grad_norm": 0.23957425355911255, + "learning_rate": 0.00010252133286999499, + "loss": 0.1675, + "step": 37870 + }, + { + "epoch": 1.4626047337735049, + "grad_norm": 1.7581062316894531, + "learning_rate": 0.00010249559185039318, + "loss": 0.241, + "step": 37880 + }, + { + "epoch": 1.4629908490675316, + "grad_norm": 1.660989761352539, + "learning_rate": 0.00010246985083079141, + "loss": 0.2099, + "step": 37890 + }, + { + "epoch": 1.4633769643615584, + "grad_norm": 1.1359142065048218, + "learning_rate": 0.00010244410981118964, + "loss": 0.3686, + "step": 37900 + }, + { + "epoch": 1.4637630796555852, + "grad_norm": 0.9489149451255798, + "learning_rate": 0.00010241836879158784, + "loss": 0.3616, + "step": 37910 + }, + { + "epoch": 1.464149194949612, + "grad_norm": 1.5209956169128418, + "learning_rate": 0.00010239262777198607, + "loss": 0.3159, + "step": 37920 + }, + { + "epoch": 1.4645353102436387, + "grad_norm": 0.848943293094635, + "learning_rate": 0.00010236688675238427, + "loss": 0.3228, + "step": 37930 + }, + { + "epoch": 1.4649214255376655, + "grad_norm": 1.1476777791976929, + "learning_rate": 0.00010234114573278248, + "loss": 0.1873, + "step": 37940 + }, + { + "epoch": 1.4653075408316925, + "grad_norm": 1.4093862771987915, + "learning_rate": 0.00010231540471318068, + "loss": 0.5554, + "step": 37950 + }, + { + "epoch": 1.465693656125719, + "grad_norm": 0.17349161207675934, + "learning_rate": 0.0001022896636935789, + "loss": 0.2992, + "step": 37960 + }, + { + "epoch": 1.466079771419746, + "grad_norm": 0.28885993361473083, + "learning_rate": 0.00010226392267397713, + "loss": 0.2618, + "step": 37970 + }, + { + "epoch": 1.4664658867137728, + "grad_norm": 1.1087830066680908, + "learning_rate": 0.00010223818165437533, + "loss": 0.2647, + "step": 37980 + }, + { + "epoch": 1.4668520020077995, + "grad_norm": 0.08338876068592072, + "learning_rate": 0.00010221244063477356, + "loss": 0.3057, + "step": 37990 + }, + { + "epoch": 1.4672381173018263, + "grad_norm": 2.159362316131592, + "learning_rate": 0.00010218669961517176, + "loss": 0.1499, + "step": 38000 + }, + { + "epoch": 1.467624232595853, + "grad_norm": 0.8207988142967224, + "learning_rate": 0.00010216095859556997, + "loss": 0.1482, + "step": 38010 + }, + { + "epoch": 1.46801034788988, + "grad_norm": 0.5458611845970154, + "learning_rate": 0.00010213521757596818, + "loss": 0.2794, + "step": 38020 + }, + { + "epoch": 1.4683964631839066, + "grad_norm": 1.6955047845840454, + "learning_rate": 0.0001021094765563664, + "loss": 0.2627, + "step": 38030 + }, + { + "epoch": 1.4687825784779336, + "grad_norm": 0.9796440601348877, + "learning_rate": 0.00010208373553676463, + "loss": 0.1924, + "step": 38040 + }, + { + "epoch": 1.4691686937719604, + "grad_norm": 0.9906508326530457, + "learning_rate": 0.00010205799451716282, + "loss": 0.1597, + "step": 38050 + }, + { + "epoch": 1.4695548090659871, + "grad_norm": 1.8590656518936157, + "learning_rate": 0.00010203225349756105, + "loss": 0.2874, + "step": 38060 + }, + { + "epoch": 1.469940924360014, + "grad_norm": 1.6403672695159912, + "learning_rate": 0.00010200651247795925, + "loss": 0.2926, + "step": 38070 + }, + { + "epoch": 1.4703270396540407, + "grad_norm": 0.4410895109176636, + "learning_rate": 0.00010198077145835746, + "loss": 0.1368, + "step": 38080 + }, + { + "epoch": 1.4707131549480674, + "grad_norm": 2.4955286979675293, + "learning_rate": 0.00010195503043875569, + "loss": 0.3195, + "step": 38090 + }, + { + "epoch": 1.4710992702420942, + "grad_norm": 1.1799029111862183, + "learning_rate": 0.00010192928941915389, + "loss": 0.1867, + "step": 38100 + }, + { + "epoch": 1.4714853855361212, + "grad_norm": 0.7959389090538025, + "learning_rate": 0.00010190354839955212, + "loss": 0.2988, + "step": 38110 + }, + { + "epoch": 1.471871500830148, + "grad_norm": 2.7750720977783203, + "learning_rate": 0.00010187780737995032, + "loss": 0.1886, + "step": 38120 + }, + { + "epoch": 1.4722576161241747, + "grad_norm": 1.5834373235702515, + "learning_rate": 0.00010185206636034854, + "loss": 0.3542, + "step": 38130 + }, + { + "epoch": 1.4726437314182015, + "grad_norm": 1.9757747650146484, + "learning_rate": 0.00010182632534074674, + "loss": 0.302, + "step": 38140 + }, + { + "epoch": 1.4730298467122283, + "grad_norm": 1.1752204895019531, + "learning_rate": 0.00010180058432114497, + "loss": 0.3535, + "step": 38150 + }, + { + "epoch": 1.473415962006255, + "grad_norm": 0.3877789378166199, + "learning_rate": 0.00010177484330154318, + "loss": 0.2508, + "step": 38160 + }, + { + "epoch": 1.4738020773002818, + "grad_norm": 0.13749545812606812, + "learning_rate": 0.00010174910228194138, + "loss": 0.2141, + "step": 38170 + }, + { + "epoch": 1.4741881925943088, + "grad_norm": 1.3663641214370728, + "learning_rate": 0.00010172336126233961, + "loss": 0.3231, + "step": 38180 + }, + { + "epoch": 1.4745743078883353, + "grad_norm": 1.6267393827438354, + "learning_rate": 0.00010169762024273781, + "loss": 0.3233, + "step": 38190 + }, + { + "epoch": 1.4749604231823623, + "grad_norm": 0.2993789315223694, + "learning_rate": 0.00010167187922313604, + "loss": 0.28, + "step": 38200 + }, + { + "epoch": 1.475346538476389, + "grad_norm": 0.16693222522735596, + "learning_rate": 0.00010164613820353424, + "loss": 0.188, + "step": 38210 + }, + { + "epoch": 1.4757326537704158, + "grad_norm": 0.6939979791641235, + "learning_rate": 0.00010162039718393246, + "loss": 0.263, + "step": 38220 + }, + { + "epoch": 1.4761187690644426, + "grad_norm": 0.37910985946655273, + "learning_rate": 0.00010159465616433068, + "loss": 0.1963, + "step": 38230 + }, + { + "epoch": 1.4765048843584694, + "grad_norm": 1.782188892364502, + "learning_rate": 0.00010156891514472888, + "loss": 0.3814, + "step": 38240 + }, + { + "epoch": 1.4768909996524964, + "grad_norm": 1.159278392791748, + "learning_rate": 0.0001015431741251271, + "loss": 0.2043, + "step": 38250 + }, + { + "epoch": 1.477277114946523, + "grad_norm": 1.09486985206604, + "learning_rate": 0.0001015174331055253, + "loss": 0.2128, + "step": 38260 + }, + { + "epoch": 1.47766323024055, + "grad_norm": 0.36655205488204956, + "learning_rate": 0.00010149169208592353, + "loss": 0.298, + "step": 38270 + }, + { + "epoch": 1.4780493455345767, + "grad_norm": 0.8908851742744446, + "learning_rate": 0.00010146595106632176, + "loss": 0.3707, + "step": 38280 + }, + { + "epoch": 1.4784354608286034, + "grad_norm": 0.251338928937912, + "learning_rate": 0.00010144021004671996, + "loss": 0.2495, + "step": 38290 + }, + { + "epoch": 1.4788215761226302, + "grad_norm": 1.0613712072372437, + "learning_rate": 0.00010141446902711817, + "loss": 0.2112, + "step": 38300 + }, + { + "epoch": 1.479207691416657, + "grad_norm": 1.459799885749817, + "learning_rate": 0.00010138872800751637, + "loss": 0.2595, + "step": 38310 + }, + { + "epoch": 1.4795938067106837, + "grad_norm": 2.6898603439331055, + "learning_rate": 0.0001013629869879146, + "loss": 0.2758, + "step": 38320 + }, + { + "epoch": 1.4799799220047105, + "grad_norm": 0.19628773629665375, + "learning_rate": 0.0001013372459683128, + "loss": 0.1843, + "step": 38330 + }, + { + "epoch": 1.4803660372987375, + "grad_norm": 2.0871078968048096, + "learning_rate": 0.00010131150494871102, + "loss": 0.1661, + "step": 38340 + }, + { + "epoch": 1.4807521525927643, + "grad_norm": 0.7689336538314819, + "learning_rate": 0.00010128576392910925, + "loss": 0.157, + "step": 38350 + }, + { + "epoch": 1.481138267886791, + "grad_norm": 1.4471644163131714, + "learning_rate": 0.00010126002290950745, + "loss": 0.2159, + "step": 38360 + }, + { + "epoch": 1.4815243831808178, + "grad_norm": 2.198559522628784, + "learning_rate": 0.00010123428188990566, + "loss": 0.5609, + "step": 38370 + }, + { + "epoch": 1.4819104984748446, + "grad_norm": 0.16012130677700043, + "learning_rate": 0.00010120854087030386, + "loss": 0.1979, + "step": 38380 + }, + { + "epoch": 1.4822966137688713, + "grad_norm": 1.0222225189208984, + "learning_rate": 0.00010118279985070209, + "loss": 0.1547, + "step": 38390 + }, + { + "epoch": 1.482682729062898, + "grad_norm": 2.7192416191101074, + "learning_rate": 0.00010115705883110032, + "loss": 0.4942, + "step": 38400 + }, + { + "epoch": 1.483068844356925, + "grad_norm": 1.887128472328186, + "learning_rate": 0.00010113131781149852, + "loss": 0.177, + "step": 38410 + }, + { + "epoch": 1.4834549596509516, + "grad_norm": 2.7628560066223145, + "learning_rate": 0.00010110557679189674, + "loss": 0.2931, + "step": 38420 + }, + { + "epoch": 1.4838410749449786, + "grad_norm": 0.3852572739124298, + "learning_rate": 0.00010107983577229494, + "loss": 0.3392, + "step": 38430 + }, + { + "epoch": 1.4842271902390054, + "grad_norm": 1.047448992729187, + "learning_rate": 0.00010105409475269316, + "loss": 0.3741, + "step": 38440 + }, + { + "epoch": 1.4846133055330322, + "grad_norm": 1.4930602312088013, + "learning_rate": 0.00010102835373309136, + "loss": 0.2564, + "step": 38450 + }, + { + "epoch": 1.484999420827059, + "grad_norm": 1.3012608289718628, + "learning_rate": 0.00010100261271348958, + "loss": 0.3376, + "step": 38460 + }, + { + "epoch": 1.4853855361210857, + "grad_norm": 2.163942337036133, + "learning_rate": 0.00010097687169388781, + "loss": 0.3548, + "step": 38470 + }, + { + "epoch": 1.4857716514151125, + "grad_norm": 1.864189624786377, + "learning_rate": 0.00010095113067428601, + "loss": 0.165, + "step": 38480 + }, + { + "epoch": 1.4861577667091392, + "grad_norm": 0.5661312341690063, + "learning_rate": 0.00010092538965468424, + "loss": 0.1764, + "step": 38490 + }, + { + "epoch": 1.4865438820031662, + "grad_norm": 0.13517481088638306, + "learning_rate": 0.00010089964863508244, + "loss": 0.5223, + "step": 38500 + }, + { + "epoch": 1.486929997297193, + "grad_norm": 0.665143609046936, + "learning_rate": 0.00010087390761548065, + "loss": 0.1943, + "step": 38510 + }, + { + "epoch": 1.4873161125912198, + "grad_norm": 1.2759610414505005, + "learning_rate": 0.00010084816659587885, + "loss": 0.3023, + "step": 38520 + }, + { + "epoch": 1.4877022278852465, + "grad_norm": 1.3209573030471802, + "learning_rate": 0.00010082242557627708, + "loss": 0.1101, + "step": 38530 + }, + { + "epoch": 1.4880883431792733, + "grad_norm": 1.2501552104949951, + "learning_rate": 0.0001007966845566753, + "loss": 0.3931, + "step": 38540 + }, + { + "epoch": 1.4884744584733, + "grad_norm": 0.6862074732780457, + "learning_rate": 0.0001007709435370735, + "loss": 0.4093, + "step": 38550 + }, + { + "epoch": 1.4888605737673268, + "grad_norm": 1.90501070022583, + "learning_rate": 0.00010074520251747173, + "loss": 0.2553, + "step": 38560 + }, + { + "epoch": 1.4892466890613538, + "grad_norm": 1.6547000408172607, + "learning_rate": 0.00010071946149786993, + "loss": 0.1558, + "step": 38570 + }, + { + "epoch": 1.4896328043553806, + "grad_norm": 0.35097751021385193, + "learning_rate": 0.00010069372047826814, + "loss": 0.2253, + "step": 38580 + }, + { + "epoch": 1.4900189196494074, + "grad_norm": 0.15141837298870087, + "learning_rate": 0.00010066797945866637, + "loss": 0.2124, + "step": 38590 + }, + { + "epoch": 1.4904050349434341, + "grad_norm": 1.7070786952972412, + "learning_rate": 0.00010064223843906457, + "loss": 0.3181, + "step": 38600 + }, + { + "epoch": 1.490791150237461, + "grad_norm": 0.5400305390357971, + "learning_rate": 0.0001006164974194628, + "loss": 0.3203, + "step": 38610 + }, + { + "epoch": 1.4911772655314877, + "grad_norm": 1.6475050449371338, + "learning_rate": 0.000100590756399861, + "loss": 0.2965, + "step": 38620 + }, + { + "epoch": 1.4915633808255144, + "grad_norm": 0.21372176706790924, + "learning_rate": 0.00010056501538025922, + "loss": 0.1456, + "step": 38630 + }, + { + "epoch": 1.4919494961195414, + "grad_norm": 0.24179309606552124, + "learning_rate": 0.00010053927436065742, + "loss": 0.2077, + "step": 38640 + }, + { + "epoch": 1.492335611413568, + "grad_norm": 1.2079945802688599, + "learning_rate": 0.00010051353334105564, + "loss": 0.1554, + "step": 38650 + }, + { + "epoch": 1.492721726707595, + "grad_norm": 1.8915836811065674, + "learning_rate": 0.00010048779232145386, + "loss": 0.3815, + "step": 38660 + }, + { + "epoch": 1.4931078420016217, + "grad_norm": 1.8128750324249268, + "learning_rate": 0.00010046205130185206, + "loss": 0.1944, + "step": 38670 + }, + { + "epoch": 1.4934939572956485, + "grad_norm": 2.0955018997192383, + "learning_rate": 0.00010043631028225029, + "loss": 0.2468, + "step": 38680 + }, + { + "epoch": 1.4938800725896753, + "grad_norm": 3.0578064918518066, + "learning_rate": 0.00010041056926264849, + "loss": 0.2186, + "step": 38690 + }, + { + "epoch": 1.494266187883702, + "grad_norm": 0.7817699909210205, + "learning_rate": 0.00010038482824304672, + "loss": 0.1154, + "step": 38700 + }, + { + "epoch": 1.4946523031777288, + "grad_norm": 6.556485176086426, + "learning_rate": 0.00010035908722344493, + "loss": 0.4752, + "step": 38710 + }, + { + "epoch": 1.4950384184717556, + "grad_norm": 2.1970055103302, + "learning_rate": 0.00010033334620384314, + "loss": 0.3234, + "step": 38720 + }, + { + "epoch": 1.4954245337657825, + "grad_norm": 1.8929531574249268, + "learning_rate": 0.00010030760518424136, + "loss": 0.1605, + "step": 38730 + }, + { + "epoch": 1.4958106490598093, + "grad_norm": 0.3992670178413391, + "learning_rate": 0.00010028186416463956, + "loss": 0.2417, + "step": 38740 + }, + { + "epoch": 1.496196764353836, + "grad_norm": 1.7508872747421265, + "learning_rate": 0.00010025612314503778, + "loss": 0.329, + "step": 38750 + }, + { + "epoch": 1.4965828796478629, + "grad_norm": 0.24950659275054932, + "learning_rate": 0.00010023038212543598, + "loss": 0.1159, + "step": 38760 + }, + { + "epoch": 1.4969689949418896, + "grad_norm": 1.5187748670578003, + "learning_rate": 0.00010020464110583421, + "loss": 0.1827, + "step": 38770 + }, + { + "epoch": 1.4973551102359164, + "grad_norm": 1.1223959922790527, + "learning_rate": 0.00010017890008623242, + "loss": 0.1935, + "step": 38780 + }, + { + "epoch": 1.4977412255299432, + "grad_norm": 0.9355156421661377, + "learning_rate": 0.00010015315906663064, + "loss": 0.2225, + "step": 38790 + }, + { + "epoch": 1.4981273408239701, + "grad_norm": 0.4834296405315399, + "learning_rate": 0.00010012741804702885, + "loss": 0.3077, + "step": 38800 + }, + { + "epoch": 1.498513456117997, + "grad_norm": 1.0983386039733887, + "learning_rate": 0.00010010167702742705, + "loss": 0.183, + "step": 38810 + }, + { + "epoch": 1.4988995714120237, + "grad_norm": 0.8350847959518433, + "learning_rate": 0.00010007593600782528, + "loss": 0.3972, + "step": 38820 + }, + { + "epoch": 1.4992856867060504, + "grad_norm": 0.8200152516365051, + "learning_rate": 0.00010005019498822348, + "loss": 0.2043, + "step": 38830 + }, + { + "epoch": 1.4996718020000772, + "grad_norm": 0.9136185050010681, + "learning_rate": 0.0001000244539686217, + "loss": 0.1727, + "step": 38840 + }, + { + "epoch": 1.500057917294104, + "grad_norm": 0.8466988205909729, + "learning_rate": 9.999871294901992e-05, + "loss": 0.2119, + "step": 38850 + }, + { + "epoch": 1.5004440325881307, + "grad_norm": 0.4185144305229187, + "learning_rate": 9.997297192941813e-05, + "loss": 0.4046, + "step": 38860 + }, + { + "epoch": 1.5008301478821577, + "grad_norm": 2.232264518737793, + "learning_rate": 9.994723090981634e-05, + "loss": 0.304, + "step": 38870 + }, + { + "epoch": 1.5012162631761843, + "grad_norm": 0.13785889744758606, + "learning_rate": 9.992148989021456e-05, + "loss": 0.1045, + "step": 38880 + }, + { + "epoch": 1.5016023784702113, + "grad_norm": 1.6270711421966553, + "learning_rate": 9.989574887061277e-05, + "loss": 0.2791, + "step": 38890 + }, + { + "epoch": 1.501988493764238, + "grad_norm": 0.08486157655715942, + "learning_rate": 9.987000785101098e-05, + "loss": 0.178, + "step": 38900 + }, + { + "epoch": 1.5023746090582648, + "grad_norm": 1.862197995185852, + "learning_rate": 9.98442668314092e-05, + "loss": 0.2461, + "step": 38910 + }, + { + "epoch": 1.5027607243522916, + "grad_norm": 2.752070903778076, + "learning_rate": 9.981852581180741e-05, + "loss": 0.4299, + "step": 38920 + }, + { + "epoch": 1.5031468396463183, + "grad_norm": 2.08542537689209, + "learning_rate": 9.979278479220562e-05, + "loss": 0.2898, + "step": 38930 + }, + { + "epoch": 1.5035329549403453, + "grad_norm": 0.8629382848739624, + "learning_rate": 9.976704377260384e-05, + "loss": 0.226, + "step": 38940 + }, + { + "epoch": 1.5039190702343719, + "grad_norm": 0.5178211331367493, + "learning_rate": 9.974130275300205e-05, + "loss": 0.2444, + "step": 38950 + }, + { + "epoch": 1.5043051855283989, + "grad_norm": 0.25908491015434265, + "learning_rate": 9.971556173340026e-05, + "loss": 0.1643, + "step": 38960 + }, + { + "epoch": 1.5046913008224254, + "grad_norm": 1.1818209886550903, + "learning_rate": 9.968982071379848e-05, + "loss": 0.3187, + "step": 38970 + }, + { + "epoch": 1.5050774161164524, + "grad_norm": 0.13186976313591003, + "learning_rate": 9.966407969419669e-05, + "loss": 0.1982, + "step": 38980 + }, + { + "epoch": 1.5054635314104792, + "grad_norm": 0.18049825727939606, + "learning_rate": 9.963833867459492e-05, + "loss": 0.1288, + "step": 38990 + }, + { + "epoch": 1.505849646704506, + "grad_norm": 0.30261853337287903, + "learning_rate": 9.961259765499312e-05, + "loss": 0.1704, + "step": 39000 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.374935773184e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d90853c06b56bd4cfca920d690278ab5a967ce5f --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb1f5b16744b8005657524508e414c6b7a7a7437d8ff1f621e2bb159cf125e27 +size 290717696 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..15a092d3b85cad7739a0ac12878ce66bd31a149a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e46acad30db0d1793166ad54c30f7696e8b81cf6ab7867280653e7b16e90c9eb +size 291241984 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..82f3a040fbd5684196456f6f571240abf9f981e1 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..5d66e7e155edeeb5d13d9ec249f00df4e5e2384c Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..68c5936abfabe5a95d6bab4738e6130e5b2dd786 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/trainer_state.json @@ -0,0 +1,27684 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.5251554114058457, + "eval_steps": 500, + "global_step": 39500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0003861152940267964, + "grad_norm": 8.003422737121582, + "learning_rate": 0.0001999768330823584, + "loss": 5.0107, + "step": 10 + }, + { + "epoch": 0.0007722305880535929, + "grad_norm": 11.965606689453125, + "learning_rate": 0.0001999510920627566, + "loss": 0.9013, + "step": 20 + }, + { + "epoch": 0.0011583458820803893, + "grad_norm": 9.783374786376953, + "learning_rate": 0.00019992535104315483, + "loss": 0.5506, + "step": 30 + }, + { + "epoch": 0.0015444611761071857, + "grad_norm": 5.720436096191406, + "learning_rate": 0.00019989961002355304, + "loss": 0.501, + "step": 40 + }, + { + "epoch": 0.001930576470133982, + "grad_norm": 11.18126392364502, + "learning_rate": 0.00019987386900395125, + "loss": 0.8176, + "step": 50 + }, + { + "epoch": 0.0023166917641607786, + "grad_norm": 8.88875675201416, + "learning_rate": 0.00019984812798434947, + "loss": 0.5404, + "step": 60 + }, + { + "epoch": 0.0027028070581875748, + "grad_norm": 2.9886066913604736, + "learning_rate": 0.00019982238696474768, + "loss": 0.5295, + "step": 70 + }, + { + "epoch": 0.0030889223522143714, + "grad_norm": 8.936307907104492, + "learning_rate": 0.0001997966459451459, + "loss": 0.7398, + "step": 80 + }, + { + "epoch": 0.0034750376462411676, + "grad_norm": 11.393534660339355, + "learning_rate": 0.0001997709049255441, + "loss": 0.6333, + "step": 90 + }, + { + "epoch": 0.003861152940267964, + "grad_norm": 8.725994110107422, + "learning_rate": 0.00019974516390594235, + "loss": 0.4689, + "step": 100 + }, + { + "epoch": 0.00424726823429476, + "grad_norm": 20.316652297973633, + "learning_rate": 0.00019971942288634053, + "loss": 0.8522, + "step": 110 + }, + { + "epoch": 0.004633383528321557, + "grad_norm": 4.232663631439209, + "learning_rate": 0.00019969368186673875, + "loss": 0.5041, + "step": 120 + }, + { + "epoch": 0.005019498822348353, + "grad_norm": 2.609255313873291, + "learning_rate": 0.00019966794084713696, + "loss": 0.7439, + "step": 130 + }, + { + "epoch": 0.0054056141163751495, + "grad_norm": 10.063919067382812, + "learning_rate": 0.00019964219982753517, + "loss": 0.6702, + "step": 140 + }, + { + "epoch": 0.005791729410401946, + "grad_norm": 5.799802303314209, + "learning_rate": 0.00019961645880793339, + "loss": 0.7356, + "step": 150 + }, + { + "epoch": 0.006177844704428743, + "grad_norm": 12.664258003234863, + "learning_rate": 0.0001995907177883316, + "loss": 0.4067, + "step": 160 + }, + { + "epoch": 0.006563959998455539, + "grad_norm": 9.42366886138916, + "learning_rate": 0.00019956497676872984, + "loss": 0.4767, + "step": 170 + }, + { + "epoch": 0.006950075292482335, + "grad_norm": 5.382272243499756, + "learning_rate": 0.00019953923574912803, + "loss": 0.536, + "step": 180 + }, + { + "epoch": 0.0073361905865091314, + "grad_norm": 9.797371864318848, + "learning_rate": 0.00019951349472952624, + "loss": 0.4735, + "step": 190 + }, + { + "epoch": 0.007722305880535928, + "grad_norm": 7.965329647064209, + "learning_rate": 0.00019948775370992445, + "loss": 0.3881, + "step": 200 + }, + { + "epoch": 0.008108421174562725, + "grad_norm": 4.075791835784912, + "learning_rate": 0.00019946201269032267, + "loss": 0.5564, + "step": 210 + }, + { + "epoch": 0.00849453646858952, + "grad_norm": 24.367305755615234, + "learning_rate": 0.0001994362716707209, + "loss": 0.9795, + "step": 220 + }, + { + "epoch": 0.008880651762616317, + "grad_norm": 9.627866744995117, + "learning_rate": 0.0001994105306511191, + "loss": 0.4528, + "step": 230 + }, + { + "epoch": 0.009266767056643114, + "grad_norm": 7.469555854797363, + "learning_rate": 0.00019938478963151733, + "loss": 0.447, + "step": 240 + }, + { + "epoch": 0.00965288235066991, + "grad_norm": 7.426730155944824, + "learning_rate": 0.00019935904861191552, + "loss": 0.6026, + "step": 250 + }, + { + "epoch": 0.010038997644696707, + "grad_norm": 6.999317169189453, + "learning_rate": 0.00019933330759231373, + "loss": 0.4962, + "step": 260 + }, + { + "epoch": 0.010425112938723502, + "grad_norm": 10.492286682128906, + "learning_rate": 0.00019930756657271194, + "loss": 0.7987, + "step": 270 + }, + { + "epoch": 0.010811228232750299, + "grad_norm": 7.079407215118408, + "learning_rate": 0.00019928182555311016, + "loss": 0.4395, + "step": 280 + }, + { + "epoch": 0.011197343526777096, + "grad_norm": 9.610014915466309, + "learning_rate": 0.0001992560845335084, + "loss": 0.748, + "step": 290 + }, + { + "epoch": 0.011583458820803891, + "grad_norm": 5.993048667907715, + "learning_rate": 0.00019923034351390658, + "loss": 0.4328, + "step": 300 + }, + { + "epoch": 0.011969574114830689, + "grad_norm": 7.336791515350342, + "learning_rate": 0.00019920460249430483, + "loss": 0.4104, + "step": 310 + }, + { + "epoch": 0.012355689408857486, + "grad_norm": 7.967221736907959, + "learning_rate": 0.000199178861474703, + "loss": 0.4662, + "step": 320 + }, + { + "epoch": 0.012741804702884281, + "grad_norm": 4.464987754821777, + "learning_rate": 0.00019915312045510125, + "loss": 0.725, + "step": 330 + }, + { + "epoch": 0.013127919996911078, + "grad_norm": 8.669449806213379, + "learning_rate": 0.00019912737943549944, + "loss": 0.4256, + "step": 340 + }, + { + "epoch": 0.013514035290937873, + "grad_norm": 4.114014148712158, + "learning_rate": 0.00019910163841589765, + "loss": 0.4477, + "step": 350 + }, + { + "epoch": 0.01390015058496467, + "grad_norm": 9.254106521606445, + "learning_rate": 0.0001990758973962959, + "loss": 0.514, + "step": 360 + }, + { + "epoch": 0.014286265878991468, + "grad_norm": 0.8039970993995667, + "learning_rate": 0.00019905015637669408, + "loss": 0.5802, + "step": 370 + }, + { + "epoch": 0.014672381173018263, + "grad_norm": 3.9931838512420654, + "learning_rate": 0.00019902441535709232, + "loss": 0.8973, + "step": 380 + }, + { + "epoch": 0.01505849646704506, + "grad_norm": 1.7645355463027954, + "learning_rate": 0.0001989986743374905, + "loss": 0.7108, + "step": 390 + }, + { + "epoch": 0.015444611761071855, + "grad_norm": 6.8542866706848145, + "learning_rate": 0.00019897293331788875, + "loss": 0.5796, + "step": 400 + }, + { + "epoch": 0.015830727055098654, + "grad_norm": 5.278103828430176, + "learning_rate": 0.00019894719229828696, + "loss": 0.3841, + "step": 410 + }, + { + "epoch": 0.01621684234912545, + "grad_norm": 9.00206184387207, + "learning_rate": 0.00019892145127868514, + "loss": 0.5891, + "step": 420 + }, + { + "epoch": 0.016602957643152245, + "grad_norm": 7.684702396392822, + "learning_rate": 0.00019889571025908339, + "loss": 0.4868, + "step": 430 + }, + { + "epoch": 0.01698907293717904, + "grad_norm": 4.198502540588379, + "learning_rate": 0.00019886996923948157, + "loss": 0.571, + "step": 440 + }, + { + "epoch": 0.01737518823120584, + "grad_norm": 7.454501628875732, + "learning_rate": 0.0001988442282198798, + "loss": 0.5133, + "step": 450 + }, + { + "epoch": 0.017761303525232634, + "grad_norm": 13.236722946166992, + "learning_rate": 0.000198818487200278, + "loss": 0.4139, + "step": 460 + }, + { + "epoch": 0.01814741881925943, + "grad_norm": 6.4592390060424805, + "learning_rate": 0.00019879274618067624, + "loss": 0.6078, + "step": 470 + }, + { + "epoch": 0.01853353411328623, + "grad_norm": 11.73417854309082, + "learning_rate": 0.00019876700516107445, + "loss": 0.5472, + "step": 480 + }, + { + "epoch": 0.018919649407313024, + "grad_norm": 2.5162808895111084, + "learning_rate": 0.00019874126414147264, + "loss": 0.6611, + "step": 490 + }, + { + "epoch": 0.01930576470133982, + "grad_norm": 4.9637837409973145, + "learning_rate": 0.00019871552312187088, + "loss": 0.6472, + "step": 500 + }, + { + "epoch": 0.019691879995366618, + "grad_norm": 11.545489311218262, + "learning_rate": 0.00019868978210226906, + "loss": 0.5304, + "step": 510 + }, + { + "epoch": 0.020077995289393413, + "grad_norm": 5.197858810424805, + "learning_rate": 0.0001986640410826673, + "loss": 0.605, + "step": 520 + }, + { + "epoch": 0.02046411058342021, + "grad_norm": 4.935055255889893, + "learning_rate": 0.0001986383000630655, + "loss": 0.6524, + "step": 530 + }, + { + "epoch": 0.020850225877447004, + "grad_norm": 5.838052749633789, + "learning_rate": 0.00019861255904346373, + "loss": 0.4957, + "step": 540 + }, + { + "epoch": 0.021236341171473803, + "grad_norm": 4.682408809661865, + "learning_rate": 0.00019858681802386194, + "loss": 0.8523, + "step": 550 + }, + { + "epoch": 0.021622456465500598, + "grad_norm": 10.720857620239258, + "learning_rate": 0.00019856107700426013, + "loss": 0.516, + "step": 560 + }, + { + "epoch": 0.022008571759527393, + "grad_norm": 6.515562534332275, + "learning_rate": 0.00019853533598465837, + "loss": 0.6095, + "step": 570 + }, + { + "epoch": 0.022394687053554192, + "grad_norm": 3.204960584640503, + "learning_rate": 0.00019850959496505656, + "loss": 0.6624, + "step": 580 + }, + { + "epoch": 0.022780802347580988, + "grad_norm": 2.305497884750366, + "learning_rate": 0.0001984838539454548, + "loss": 0.5986, + "step": 590 + }, + { + "epoch": 0.023166917641607783, + "grad_norm": 13.07105541229248, + "learning_rate": 0.000198458112925853, + "loss": 0.337, + "step": 600 + }, + { + "epoch": 0.023553032935634582, + "grad_norm": 3.1491329669952393, + "learning_rate": 0.00019843237190625122, + "loss": 0.5466, + "step": 610 + }, + { + "epoch": 0.023939148229661377, + "grad_norm": 1.92014479637146, + "learning_rate": 0.00019840663088664944, + "loss": 0.5525, + "step": 620 + }, + { + "epoch": 0.024325263523688172, + "grad_norm": 2.206550121307373, + "learning_rate": 0.00019838088986704762, + "loss": 0.5069, + "step": 630 + }, + { + "epoch": 0.02471137881771497, + "grad_norm": 2.386288642883301, + "learning_rate": 0.00019835514884744586, + "loss": 0.3749, + "step": 640 + }, + { + "epoch": 0.025097494111741767, + "grad_norm": 7.393959045410156, + "learning_rate": 0.00019832940782784405, + "loss": 0.7144, + "step": 650 + }, + { + "epoch": 0.025483609405768562, + "grad_norm": 5.7293171882629395, + "learning_rate": 0.0001983036668082423, + "loss": 0.5052, + "step": 660 + }, + { + "epoch": 0.025869724699795357, + "grad_norm": 6.440220832824707, + "learning_rate": 0.0001982779257886405, + "loss": 0.4343, + "step": 670 + }, + { + "epoch": 0.026255839993822156, + "grad_norm": 0.8553487658500671, + "learning_rate": 0.00019825218476903872, + "loss": 0.7562, + "step": 680 + }, + { + "epoch": 0.02664195528784895, + "grad_norm": 3.762784719467163, + "learning_rate": 0.00019822644374943693, + "loss": 0.5593, + "step": 690 + }, + { + "epoch": 0.027028070581875747, + "grad_norm": 11.674392700195312, + "learning_rate": 0.00019820070272983512, + "loss": 0.6069, + "step": 700 + }, + { + "epoch": 0.027414185875902546, + "grad_norm": 8.631232261657715, + "learning_rate": 0.00019817496171023336, + "loss": 0.3584, + "step": 710 + }, + { + "epoch": 0.02780030116992934, + "grad_norm": 5.7163920402526855, + "learning_rate": 0.00019814922069063157, + "loss": 0.5563, + "step": 720 + }, + { + "epoch": 0.028186416463956136, + "grad_norm": 8.186172485351562, + "learning_rate": 0.00019812347967102978, + "loss": 0.5289, + "step": 730 + }, + { + "epoch": 0.028572531757982935, + "grad_norm": 7.287814140319824, + "learning_rate": 0.000198097738651428, + "loss": 0.4543, + "step": 740 + }, + { + "epoch": 0.02895864705200973, + "grad_norm": 6.621245384216309, + "learning_rate": 0.0001980719976318262, + "loss": 0.3244, + "step": 750 + }, + { + "epoch": 0.029344762346036526, + "grad_norm": 3.5209403038024902, + "learning_rate": 0.00019804625661222442, + "loss": 0.5385, + "step": 760 + }, + { + "epoch": 0.029730877640063325, + "grad_norm": 2.562343120574951, + "learning_rate": 0.00019802051559262264, + "loss": 0.4868, + "step": 770 + }, + { + "epoch": 0.03011699293409012, + "grad_norm": 7.782780647277832, + "learning_rate": 0.00019799477457302085, + "loss": 0.5682, + "step": 780 + }, + { + "epoch": 0.030503108228116915, + "grad_norm": 8.173531532287598, + "learning_rate": 0.00019796903355341906, + "loss": 0.3557, + "step": 790 + }, + { + "epoch": 0.03088922352214371, + "grad_norm": 4.502675533294678, + "learning_rate": 0.00019794329253381728, + "loss": 0.517, + "step": 800 + }, + { + "epoch": 0.031275338816170506, + "grad_norm": 6.314894199371338, + "learning_rate": 0.0001979175515142155, + "loss": 0.6392, + "step": 810 + }, + { + "epoch": 0.03166145411019731, + "grad_norm": 7.054763317108154, + "learning_rate": 0.0001978918104946137, + "loss": 0.5823, + "step": 820 + }, + { + "epoch": 0.032047569404224104, + "grad_norm": 1.7847551107406616, + "learning_rate": 0.00019786606947501192, + "loss": 0.4495, + "step": 830 + }, + { + "epoch": 0.0324336846982509, + "grad_norm": 5.268612861633301, + "learning_rate": 0.00019784032845541013, + "loss": 0.4379, + "step": 840 + }, + { + "epoch": 0.032819799992277694, + "grad_norm": 4.834717273712158, + "learning_rate": 0.00019781458743580834, + "loss": 0.5379, + "step": 850 + }, + { + "epoch": 0.03320591528630449, + "grad_norm": 6.077203273773193, + "learning_rate": 0.00019778884641620656, + "loss": 0.5666, + "step": 860 + }, + { + "epoch": 0.033592030580331285, + "grad_norm": 0.9583851099014282, + "learning_rate": 0.00019776310539660477, + "loss": 0.8146, + "step": 870 + }, + { + "epoch": 0.03397814587435808, + "grad_norm": 0.8457469940185547, + "learning_rate": 0.00019773736437700298, + "loss": 0.3497, + "step": 880 + }, + { + "epoch": 0.03436426116838488, + "grad_norm": 5.251153945922852, + "learning_rate": 0.0001977116233574012, + "loss": 0.6299, + "step": 890 + }, + { + "epoch": 0.03475037646241168, + "grad_norm": 4.057605266571045, + "learning_rate": 0.0001976858823377994, + "loss": 0.5829, + "step": 900 + }, + { + "epoch": 0.03513649175643847, + "grad_norm": 7.625199794769287, + "learning_rate": 0.00019766014131819762, + "loss": 0.452, + "step": 910 + }, + { + "epoch": 0.03552260705046527, + "grad_norm": 4.618866443634033, + "learning_rate": 0.00019763440029859584, + "loss": 0.5693, + "step": 920 + }, + { + "epoch": 0.035908722344492064, + "grad_norm": 8.480955123901367, + "learning_rate": 0.00019760865927899405, + "loss": 0.4894, + "step": 930 + }, + { + "epoch": 0.03629483763851886, + "grad_norm": 6.537581920623779, + "learning_rate": 0.00019758291825939226, + "loss": 0.7396, + "step": 940 + }, + { + "epoch": 0.03668095293254566, + "grad_norm": 8.093205451965332, + "learning_rate": 0.00019755717723979048, + "loss": 0.3702, + "step": 950 + }, + { + "epoch": 0.03706706822657246, + "grad_norm": 2.523141860961914, + "learning_rate": 0.0001975314362201887, + "loss": 0.3746, + "step": 960 + }, + { + "epoch": 0.03745318352059925, + "grad_norm": 1.2707194089889526, + "learning_rate": 0.0001975056952005869, + "loss": 0.3211, + "step": 970 + }, + { + "epoch": 0.03783929881462605, + "grad_norm": 2.818798780441284, + "learning_rate": 0.00019747995418098512, + "loss": 0.6594, + "step": 980 + }, + { + "epoch": 0.03822541410865284, + "grad_norm": 7.373154640197754, + "learning_rate": 0.00019745421316138333, + "loss": 0.7825, + "step": 990 + }, + { + "epoch": 0.03861152940267964, + "grad_norm": 2.8925669193267822, + "learning_rate": 0.00019742847214178154, + "loss": 0.4847, + "step": 1000 + }, + { + "epoch": 0.038997644696706434, + "grad_norm": 10.87833023071289, + "learning_rate": 0.00019740273112217976, + "loss": 0.5098, + "step": 1010 + }, + { + "epoch": 0.039383759990733236, + "grad_norm": 3.8262317180633545, + "learning_rate": 0.00019737699010257797, + "loss": 0.5168, + "step": 1020 + }, + { + "epoch": 0.03976987528476003, + "grad_norm": 2.5567004680633545, + "learning_rate": 0.00019735124908297618, + "loss": 0.5597, + "step": 1030 + }, + { + "epoch": 0.04015599057878683, + "grad_norm": 4.385695457458496, + "learning_rate": 0.0001973255080633744, + "loss": 0.3187, + "step": 1040 + }, + { + "epoch": 0.04054210587281362, + "grad_norm": 1.8186907768249512, + "learning_rate": 0.0001972997670437726, + "loss": 0.6274, + "step": 1050 + }, + { + "epoch": 0.04092822116684042, + "grad_norm": 7.446481704711914, + "learning_rate": 0.00019727402602417082, + "loss": 0.4365, + "step": 1060 + }, + { + "epoch": 0.04131433646086721, + "grad_norm": 8.973576545715332, + "learning_rate": 0.00019724828500456904, + "loss": 0.6518, + "step": 1070 + }, + { + "epoch": 0.04170045175489401, + "grad_norm": 4.984101295471191, + "learning_rate": 0.00019722254398496725, + "loss": 0.4694, + "step": 1080 + }, + { + "epoch": 0.04208656704892081, + "grad_norm": 8.2625093460083, + "learning_rate": 0.00019719680296536546, + "loss": 0.5532, + "step": 1090 + }, + { + "epoch": 0.042472682342947606, + "grad_norm": 2.2425265312194824, + "learning_rate": 0.0001971710619457637, + "loss": 0.4011, + "step": 1100 + }, + { + "epoch": 0.0428587976369744, + "grad_norm": 1.4552969932556152, + "learning_rate": 0.0001971453209261619, + "loss": 0.396, + "step": 1110 + }, + { + "epoch": 0.043244912931001196, + "grad_norm": 2.0675857067108154, + "learning_rate": 0.0001971195799065601, + "loss": 0.3727, + "step": 1120 + }, + { + "epoch": 0.04363102822502799, + "grad_norm": 5.24460506439209, + "learning_rate": 0.00019709383888695832, + "loss": 0.5016, + "step": 1130 + }, + { + "epoch": 0.04401714351905479, + "grad_norm": 5.524955749511719, + "learning_rate": 0.00019706809786735653, + "loss": 0.5866, + "step": 1140 + }, + { + "epoch": 0.04440325881308159, + "grad_norm": 4.2749199867248535, + "learning_rate": 0.00019704235684775474, + "loss": 0.4591, + "step": 1150 + }, + { + "epoch": 0.044789374107108385, + "grad_norm": 3.460395336151123, + "learning_rate": 0.00019701661582815296, + "loss": 0.5275, + "step": 1160 + }, + { + "epoch": 0.04517548940113518, + "grad_norm": 1.3979772329330444, + "learning_rate": 0.0001969908748085512, + "loss": 0.3375, + "step": 1170 + }, + { + "epoch": 0.045561604695161975, + "grad_norm": 2.29923939704895, + "learning_rate": 0.00019696513378894938, + "loss": 0.5683, + "step": 1180 + }, + { + "epoch": 0.04594771998918877, + "grad_norm": 3.211496353149414, + "learning_rate": 0.0001969393927693476, + "loss": 0.7122, + "step": 1190 + }, + { + "epoch": 0.046333835283215566, + "grad_norm": 4.18447208404541, + "learning_rate": 0.0001969136517497458, + "loss": 0.5149, + "step": 1200 + }, + { + "epoch": 0.04671995057724236, + "grad_norm": 14.650918960571289, + "learning_rate": 0.00019688791073014402, + "loss": 0.6384, + "step": 1210 + }, + { + "epoch": 0.047106065871269164, + "grad_norm": 4.956239700317383, + "learning_rate": 0.00019686216971054226, + "loss": 0.3602, + "step": 1220 + }, + { + "epoch": 0.04749218116529596, + "grad_norm": 4.0486860275268555, + "learning_rate": 0.00019683642869094045, + "loss": 0.7719, + "step": 1230 + }, + { + "epoch": 0.047878296459322754, + "grad_norm": 4.587133407592773, + "learning_rate": 0.0001968106876713387, + "loss": 0.4079, + "step": 1240 + }, + { + "epoch": 0.04826441175334955, + "grad_norm": 0.7830008268356323, + "learning_rate": 0.00019678494665173688, + "loss": 0.5841, + "step": 1250 + }, + { + "epoch": 0.048650527047376345, + "grad_norm": 5.378068447113037, + "learning_rate": 0.0001967592056321351, + "loss": 0.3226, + "step": 1260 + }, + { + "epoch": 0.04903664234140314, + "grad_norm": 4.002605438232422, + "learning_rate": 0.0001967334646125333, + "loss": 0.4411, + "step": 1270 + }, + { + "epoch": 0.04942275763542994, + "grad_norm": 4.695134162902832, + "learning_rate": 0.00019670772359293152, + "loss": 0.415, + "step": 1280 + }, + { + "epoch": 0.04980887292945674, + "grad_norm": 6.046143054962158, + "learning_rate": 0.00019668198257332976, + "loss": 0.463, + "step": 1290 + }, + { + "epoch": 0.05019498822348353, + "grad_norm": 1.8300361633300781, + "learning_rate": 0.00019665624155372794, + "loss": 0.408, + "step": 1300 + }, + { + "epoch": 0.05058110351751033, + "grad_norm": 5.80141544342041, + "learning_rate": 0.00019663050053412618, + "loss": 0.4481, + "step": 1310 + }, + { + "epoch": 0.050967218811537124, + "grad_norm": 4.103593349456787, + "learning_rate": 0.00019660475951452437, + "loss": 0.3054, + "step": 1320 + }, + { + "epoch": 0.05135333410556392, + "grad_norm": 9.129929542541504, + "learning_rate": 0.00019657901849492258, + "loss": 0.5554, + "step": 1330 + }, + { + "epoch": 0.051739449399590715, + "grad_norm": 6.979663372039795, + "learning_rate": 0.0001965532774753208, + "loss": 0.388, + "step": 1340 + }, + { + "epoch": 0.05212556469361752, + "grad_norm": 6.329915523529053, + "learning_rate": 0.000196527536455719, + "loss": 0.8894, + "step": 1350 + }, + { + "epoch": 0.05251167998764431, + "grad_norm": 0.2686227262020111, + "learning_rate": 0.00019650179543611725, + "loss": 0.5032, + "step": 1360 + }, + { + "epoch": 0.05289779528167111, + "grad_norm": 4.818896770477295, + "learning_rate": 0.00019647605441651544, + "loss": 0.3417, + "step": 1370 + }, + { + "epoch": 0.0532839105756979, + "grad_norm": 6.252008438110352, + "learning_rate": 0.00019645031339691368, + "loss": 0.398, + "step": 1380 + }, + { + "epoch": 0.0536700258697247, + "grad_norm": 1.2734620571136475, + "learning_rate": 0.00019642457237731186, + "loss": 0.5369, + "step": 1390 + }, + { + "epoch": 0.054056141163751494, + "grad_norm": 6.511690616607666, + "learning_rate": 0.00019639883135771008, + "loss": 0.4646, + "step": 1400 + }, + { + "epoch": 0.054442256457778296, + "grad_norm": 3.2352371215820312, + "learning_rate": 0.00019637309033810832, + "loss": 0.3805, + "step": 1410 + }, + { + "epoch": 0.05482837175180509, + "grad_norm": 1.0574132204055786, + "learning_rate": 0.0001963473493185065, + "loss": 0.4572, + "step": 1420 + }, + { + "epoch": 0.055214487045831886, + "grad_norm": 2.8525452613830566, + "learning_rate": 0.00019632160829890474, + "loss": 0.4417, + "step": 1430 + }, + { + "epoch": 0.05560060233985868, + "grad_norm": 3.588179111480713, + "learning_rate": 0.00019629586727930293, + "loss": 0.6214, + "step": 1440 + }, + { + "epoch": 0.05598671763388548, + "grad_norm": 3.969320058822632, + "learning_rate": 0.00019627012625970117, + "loss": 0.6114, + "step": 1450 + }, + { + "epoch": 0.05637283292791227, + "grad_norm": 3.465053081512451, + "learning_rate": 0.00019624438524009936, + "loss": 0.6066, + "step": 1460 + }, + { + "epoch": 0.05675894822193907, + "grad_norm": 3.5419201850891113, + "learning_rate": 0.00019621864422049757, + "loss": 0.3906, + "step": 1470 + }, + { + "epoch": 0.05714506351596587, + "grad_norm": 3.4580233097076416, + "learning_rate": 0.0001961929032008958, + "loss": 0.9283, + "step": 1480 + }, + { + "epoch": 0.057531178809992665, + "grad_norm": 4.222144603729248, + "learning_rate": 0.000196167162181294, + "loss": 0.4225, + "step": 1490 + }, + { + "epoch": 0.05791729410401946, + "grad_norm": 0.8072681427001953, + "learning_rate": 0.00019614142116169224, + "loss": 0.5012, + "step": 1500 + }, + { + "epoch": 0.058303409398046256, + "grad_norm": 2.827258348464966, + "learning_rate": 0.00019611568014209042, + "loss": 0.4333, + "step": 1510 + }, + { + "epoch": 0.05868952469207305, + "grad_norm": 1.3494776487350464, + "learning_rate": 0.00019608993912248866, + "loss": 0.3019, + "step": 1520 + }, + { + "epoch": 0.05907563998609985, + "grad_norm": 4.3279900550842285, + "learning_rate": 0.00019606419810288688, + "loss": 0.4807, + "step": 1530 + }, + { + "epoch": 0.05946175528012665, + "grad_norm": 3.8996474742889404, + "learning_rate": 0.0001960384570832851, + "loss": 0.4876, + "step": 1540 + }, + { + "epoch": 0.059847870574153444, + "grad_norm": 5.255978584289551, + "learning_rate": 0.0001960127160636833, + "loss": 0.4661, + "step": 1550 + }, + { + "epoch": 0.06023398586818024, + "grad_norm": 5.172120094299316, + "learning_rate": 0.0001959869750440815, + "loss": 0.4885, + "step": 1560 + }, + { + "epoch": 0.060620101162207035, + "grad_norm": 5.385959625244141, + "learning_rate": 0.00019596123402447973, + "loss": 0.2995, + "step": 1570 + }, + { + "epoch": 0.06100621645623383, + "grad_norm": 3.9922871589660645, + "learning_rate": 0.00019593549300487792, + "loss": 0.4568, + "step": 1580 + }, + { + "epoch": 0.061392331750260626, + "grad_norm": 6.048642158508301, + "learning_rate": 0.00019590975198527616, + "loss": 0.4649, + "step": 1590 + }, + { + "epoch": 0.06177844704428742, + "grad_norm": 1.0315563678741455, + "learning_rate": 0.00019588401096567437, + "loss": 0.5175, + "step": 1600 + }, + { + "epoch": 0.06216456233831422, + "grad_norm": 12.403678894042969, + "learning_rate": 0.00019585826994607258, + "loss": 0.523, + "step": 1610 + }, + { + "epoch": 0.06255067763234101, + "grad_norm": 6.127188205718994, + "learning_rate": 0.0001958325289264708, + "loss": 0.5861, + "step": 1620 + }, + { + "epoch": 0.06293679292636781, + "grad_norm": 6.398592948913574, + "learning_rate": 0.00019580678790686898, + "loss": 0.7471, + "step": 1630 + }, + { + "epoch": 0.06332290822039462, + "grad_norm": 4.127200603485107, + "learning_rate": 0.00019578104688726722, + "loss": 0.4921, + "step": 1640 + }, + { + "epoch": 0.06370902351442141, + "grad_norm": 4.601541042327881, + "learning_rate": 0.0001957553058676654, + "loss": 0.3682, + "step": 1650 + }, + { + "epoch": 0.06409513880844821, + "grad_norm": 6.32781457901001, + "learning_rate": 0.00019572956484806365, + "loss": 0.3748, + "step": 1660 + }, + { + "epoch": 0.064481254102475, + "grad_norm": 3.7280173301696777, + "learning_rate": 0.00019570382382846186, + "loss": 0.5912, + "step": 1670 + }, + { + "epoch": 0.0648673693965018, + "grad_norm": 6.7821946144104, + "learning_rate": 0.00019567808280886008, + "loss": 0.4073, + "step": 1680 + }, + { + "epoch": 0.0652534846905286, + "grad_norm": 1.4645791053771973, + "learning_rate": 0.0001956523417892583, + "loss": 0.7164, + "step": 1690 + }, + { + "epoch": 0.06563959998455539, + "grad_norm": 2.367361545562744, + "learning_rate": 0.00019562660076965648, + "loss": 0.3859, + "step": 1700 + }, + { + "epoch": 0.06602571527858218, + "grad_norm": 2.198493480682373, + "learning_rate": 0.00019560085975005472, + "loss": 0.4928, + "step": 1710 + }, + { + "epoch": 0.06641183057260898, + "grad_norm": 1.882567048072815, + "learning_rate": 0.00019557511873045293, + "loss": 0.5861, + "step": 1720 + }, + { + "epoch": 0.06679794586663577, + "grad_norm": 6.324089527130127, + "learning_rate": 0.00019554937771085114, + "loss": 0.6249, + "step": 1730 + }, + { + "epoch": 0.06718406116066257, + "grad_norm": 4.283392906188965, + "learning_rate": 0.00019552363669124936, + "loss": 0.5403, + "step": 1740 + }, + { + "epoch": 0.06757017645468937, + "grad_norm": 4.464428424835205, + "learning_rate": 0.00019549789567164757, + "loss": 0.5815, + "step": 1750 + }, + { + "epoch": 0.06795629174871616, + "grad_norm": 0.32923218607902527, + "learning_rate": 0.00019547215465204578, + "loss": 0.3791, + "step": 1760 + }, + { + "epoch": 0.06834240704274297, + "grad_norm": 5.255763053894043, + "learning_rate": 0.00019544641363244397, + "loss": 0.4252, + "step": 1770 + }, + { + "epoch": 0.06872852233676977, + "grad_norm": 2.1615116596221924, + "learning_rate": 0.0001954206726128422, + "loss": 0.4457, + "step": 1780 + }, + { + "epoch": 0.06911463763079656, + "grad_norm": 2.0209217071533203, + "learning_rate": 0.00019539493159324042, + "loss": 0.4377, + "step": 1790 + }, + { + "epoch": 0.06950075292482336, + "grad_norm": 8.12317943572998, + "learning_rate": 0.00019536919057363864, + "loss": 0.4025, + "step": 1800 + }, + { + "epoch": 0.06988686821885015, + "grad_norm": 1.7034660577774048, + "learning_rate": 0.00019534344955403685, + "loss": 0.2915, + "step": 1810 + }, + { + "epoch": 0.07027298351287695, + "grad_norm": 4.640942096710205, + "learning_rate": 0.00019531770853443506, + "loss": 0.6588, + "step": 1820 + }, + { + "epoch": 0.07065909880690374, + "grad_norm": 4.202883243560791, + "learning_rate": 0.00019529196751483328, + "loss": 0.4442, + "step": 1830 + }, + { + "epoch": 0.07104521410093054, + "grad_norm": 3.26898193359375, + "learning_rate": 0.00019526622649523146, + "loss": 0.5065, + "step": 1840 + }, + { + "epoch": 0.07143132939495733, + "grad_norm": 8.189995765686035, + "learning_rate": 0.0001952404854756297, + "loss": 0.5258, + "step": 1850 + }, + { + "epoch": 0.07181744468898413, + "grad_norm": 3.2618284225463867, + "learning_rate": 0.00019521474445602792, + "loss": 0.5037, + "step": 1860 + }, + { + "epoch": 0.07220355998301092, + "grad_norm": 2.168548583984375, + "learning_rate": 0.00019518900343642613, + "loss": 0.4887, + "step": 1870 + }, + { + "epoch": 0.07258967527703772, + "grad_norm": 2.2029404640197754, + "learning_rate": 0.00019516326241682434, + "loss": 0.4646, + "step": 1880 + }, + { + "epoch": 0.07297579057106451, + "grad_norm": 1.561713695526123, + "learning_rate": 0.00019513752139722256, + "loss": 0.432, + "step": 1890 + }, + { + "epoch": 0.07336190586509132, + "grad_norm": 3.428372621536255, + "learning_rate": 0.00019511178037762077, + "loss": 0.4124, + "step": 1900 + }, + { + "epoch": 0.07374802115911812, + "grad_norm": 5.6706671714782715, + "learning_rate": 0.00019508603935801898, + "loss": 0.4431, + "step": 1910 + }, + { + "epoch": 0.07413413645314491, + "grad_norm": 8.872734069824219, + "learning_rate": 0.0001950602983384172, + "loss": 0.7001, + "step": 1920 + }, + { + "epoch": 0.07452025174717171, + "grad_norm": 1.6821974515914917, + "learning_rate": 0.0001950345573188154, + "loss": 0.4204, + "step": 1930 + }, + { + "epoch": 0.0749063670411985, + "grad_norm": 3.117480993270874, + "learning_rate": 0.00019500881629921362, + "loss": 0.3748, + "step": 1940 + }, + { + "epoch": 0.0752924823352253, + "grad_norm": 0.8384984731674194, + "learning_rate": 0.00019498307527961184, + "loss": 0.2636, + "step": 1950 + }, + { + "epoch": 0.0756785976292521, + "grad_norm": 2.8956708908081055, + "learning_rate": 0.00019495733426001005, + "loss": 0.4514, + "step": 1960 + }, + { + "epoch": 0.07606471292327889, + "grad_norm": 11.233087539672852, + "learning_rate": 0.00019493159324040826, + "loss": 0.4002, + "step": 1970 + }, + { + "epoch": 0.07645082821730569, + "grad_norm": 4.066893100738525, + "learning_rate": 0.00019490585222080648, + "loss": 0.4449, + "step": 1980 + }, + { + "epoch": 0.07683694351133248, + "grad_norm": 4.854077339172363, + "learning_rate": 0.0001948801112012047, + "loss": 0.4961, + "step": 1990 + }, + { + "epoch": 0.07722305880535928, + "grad_norm": 2.5722827911376953, + "learning_rate": 0.0001948543701816029, + "loss": 0.3743, + "step": 2000 + }, + { + "epoch": 0.07760917409938607, + "grad_norm": 5.842077255249023, + "learning_rate": 0.00019482862916200112, + "loss": 0.2906, + "step": 2010 + }, + { + "epoch": 0.07799528939341287, + "grad_norm": 6.163092136383057, + "learning_rate": 0.00019480288814239933, + "loss": 0.4374, + "step": 2020 + }, + { + "epoch": 0.07838140468743968, + "grad_norm": 4.589334487915039, + "learning_rate": 0.00019477714712279754, + "loss": 0.484, + "step": 2030 + }, + { + "epoch": 0.07876751998146647, + "grad_norm": 6.951212406158447, + "learning_rate": 0.00019475140610319576, + "loss": 0.5767, + "step": 2040 + }, + { + "epoch": 0.07915363527549327, + "grad_norm": 3.3662521839141846, + "learning_rate": 0.00019472566508359397, + "loss": 0.5566, + "step": 2050 + }, + { + "epoch": 0.07953975056952006, + "grad_norm": 1.6602391004562378, + "learning_rate": 0.00019469992406399218, + "loss": 0.1436, + "step": 2060 + }, + { + "epoch": 0.07992586586354686, + "grad_norm": 6.451857089996338, + "learning_rate": 0.0001946741830443904, + "loss": 0.3778, + "step": 2070 + }, + { + "epoch": 0.08031198115757365, + "grad_norm": 3.7249560356140137, + "learning_rate": 0.0001946484420247886, + "loss": 0.5391, + "step": 2080 + }, + { + "epoch": 0.08069809645160045, + "grad_norm": 4.138098239898682, + "learning_rate": 0.00019462270100518682, + "loss": 0.3598, + "step": 2090 + }, + { + "epoch": 0.08108421174562724, + "grad_norm": 3.224778175354004, + "learning_rate": 0.00019459695998558504, + "loss": 0.2967, + "step": 2100 + }, + { + "epoch": 0.08147032703965404, + "grad_norm": 1.3951358795166016, + "learning_rate": 0.00019457121896598325, + "loss": 0.2698, + "step": 2110 + }, + { + "epoch": 0.08185644233368083, + "grad_norm": 5.956802845001221, + "learning_rate": 0.00019454547794638146, + "loss": 0.451, + "step": 2120 + }, + { + "epoch": 0.08224255762770763, + "grad_norm": 3.456360101699829, + "learning_rate": 0.00019451973692677968, + "loss": 0.3365, + "step": 2130 + }, + { + "epoch": 0.08262867292173443, + "grad_norm": 2.9433653354644775, + "learning_rate": 0.0001944939959071779, + "loss": 0.4424, + "step": 2140 + }, + { + "epoch": 0.08301478821576122, + "grad_norm": 3.136000871658325, + "learning_rate": 0.0001944682548875761, + "loss": 0.6224, + "step": 2150 + }, + { + "epoch": 0.08340090350978802, + "grad_norm": 2.669084310531616, + "learning_rate": 0.00019444251386797432, + "loss": 0.4435, + "step": 2160 + }, + { + "epoch": 0.08378701880381483, + "grad_norm": 4.573731899261475, + "learning_rate": 0.00019441677284837253, + "loss": 0.5555, + "step": 2170 + }, + { + "epoch": 0.08417313409784162, + "grad_norm": 6.354156017303467, + "learning_rate": 0.00019439103182877074, + "loss": 0.4232, + "step": 2180 + }, + { + "epoch": 0.08455924939186842, + "grad_norm": 2.993691921234131, + "learning_rate": 0.00019436529080916895, + "loss": 0.51, + "step": 2190 + }, + { + "epoch": 0.08494536468589521, + "grad_norm": 3.6496782302856445, + "learning_rate": 0.00019433954978956717, + "loss": 0.4031, + "step": 2200 + }, + { + "epoch": 0.085331479979922, + "grad_norm": 1.9039051532745361, + "learning_rate": 0.00019431380876996538, + "loss": 0.4407, + "step": 2210 + }, + { + "epoch": 0.0857175952739488, + "grad_norm": 2.3211915493011475, + "learning_rate": 0.00019428806775036362, + "loss": 0.4057, + "step": 2220 + }, + { + "epoch": 0.0861037105679756, + "grad_norm": 4.883905410766602, + "learning_rate": 0.0001942623267307618, + "loss": 0.6223, + "step": 2230 + }, + { + "epoch": 0.08648982586200239, + "grad_norm": 2.164484977722168, + "learning_rate": 0.00019423658571116005, + "loss": 0.3141, + "step": 2240 + }, + { + "epoch": 0.08687594115602919, + "grad_norm": 2.2078909873962402, + "learning_rate": 0.00019421084469155823, + "loss": 0.3523, + "step": 2250 + }, + { + "epoch": 0.08726205645005598, + "grad_norm": 1.0987967252731323, + "learning_rate": 0.00019418510367195645, + "loss": 0.4013, + "step": 2260 + }, + { + "epoch": 0.08764817174408278, + "grad_norm": 2.418327569961548, + "learning_rate": 0.00019415936265235466, + "loss": 0.581, + "step": 2270 + }, + { + "epoch": 0.08803428703810957, + "grad_norm": 4.029652118682861, + "learning_rate": 0.00019413362163275287, + "loss": 0.5244, + "step": 2280 + }, + { + "epoch": 0.08842040233213637, + "grad_norm": 3.4661777019500732, + "learning_rate": 0.00019410788061315112, + "loss": 0.4531, + "step": 2290 + }, + { + "epoch": 0.08880651762616318, + "grad_norm": 1.0475856065750122, + "learning_rate": 0.0001940821395935493, + "loss": 0.4362, + "step": 2300 + }, + { + "epoch": 0.08919263292018997, + "grad_norm": 4.021854400634766, + "learning_rate": 0.00019405639857394754, + "loss": 0.4532, + "step": 2310 + }, + { + "epoch": 0.08957874821421677, + "grad_norm": 1.836438536643982, + "learning_rate": 0.00019403065755434573, + "loss": 0.6858, + "step": 2320 + }, + { + "epoch": 0.08996486350824356, + "grad_norm": 2.5251567363739014, + "learning_rate": 0.00019400491653474394, + "loss": 0.3619, + "step": 2330 + }, + { + "epoch": 0.09035097880227036, + "grad_norm": 3.067208766937256, + "learning_rate": 0.00019397917551514215, + "loss": 0.7376, + "step": 2340 + }, + { + "epoch": 0.09073709409629716, + "grad_norm": 0.9124518036842346, + "learning_rate": 0.00019395343449554037, + "loss": 0.4193, + "step": 2350 + }, + { + "epoch": 0.09112320939032395, + "grad_norm": 3.8170812129974365, + "learning_rate": 0.0001939276934759386, + "loss": 0.5393, + "step": 2360 + }, + { + "epoch": 0.09150932468435075, + "grad_norm": 8.19250202178955, + "learning_rate": 0.0001939019524563368, + "loss": 0.424, + "step": 2370 + }, + { + "epoch": 0.09189543997837754, + "grad_norm": 4.459112167358398, + "learning_rate": 0.00019387621143673503, + "loss": 0.3278, + "step": 2380 + }, + { + "epoch": 0.09228155527240434, + "grad_norm": 5.578339576721191, + "learning_rate": 0.00019385047041713322, + "loss": 0.5223, + "step": 2390 + }, + { + "epoch": 0.09266767056643113, + "grad_norm": 1.3707878589630127, + "learning_rate": 0.00019382472939753143, + "loss": 0.3004, + "step": 2400 + }, + { + "epoch": 0.09305378586045793, + "grad_norm": 5.0041184425354, + "learning_rate": 0.00019379898837792967, + "loss": 0.4378, + "step": 2410 + }, + { + "epoch": 0.09343990115448472, + "grad_norm": 5.668384552001953, + "learning_rate": 0.00019377324735832786, + "loss": 0.499, + "step": 2420 + }, + { + "epoch": 0.09382601644851153, + "grad_norm": 5.605838775634766, + "learning_rate": 0.0001937475063387261, + "loss": 0.5642, + "step": 2430 + }, + { + "epoch": 0.09421213174253833, + "grad_norm": 5.055904865264893, + "learning_rate": 0.0001937217653191243, + "loss": 0.6225, + "step": 2440 + }, + { + "epoch": 0.09459824703656512, + "grad_norm": 3.1779348850250244, + "learning_rate": 0.00019369602429952253, + "loss": 0.3673, + "step": 2450 + }, + { + "epoch": 0.09498436233059192, + "grad_norm": 2.540269136428833, + "learning_rate": 0.00019367028327992071, + "loss": 0.3499, + "step": 2460 + }, + { + "epoch": 0.09537047762461871, + "grad_norm": 2.284114122390747, + "learning_rate": 0.00019364454226031893, + "loss": 0.5405, + "step": 2470 + }, + { + "epoch": 0.09575659291864551, + "grad_norm": 6.752682685852051, + "learning_rate": 0.00019361880124071717, + "loss": 0.5594, + "step": 2480 + }, + { + "epoch": 0.0961427082126723, + "grad_norm": 2.5659310817718506, + "learning_rate": 0.00019359306022111535, + "loss": 0.5872, + "step": 2490 + }, + { + "epoch": 0.0965288235066991, + "grad_norm": 4.503110885620117, + "learning_rate": 0.0001935673192015136, + "loss": 0.5171, + "step": 2500 + }, + { + "epoch": 0.0969149388007259, + "grad_norm": 1.7715080976486206, + "learning_rate": 0.00019354157818191178, + "loss": 0.6131, + "step": 2510 + }, + { + "epoch": 0.09730105409475269, + "grad_norm": 4.479000568389893, + "learning_rate": 0.00019351583716231002, + "loss": 0.4396, + "step": 2520 + }, + { + "epoch": 0.09768716938877949, + "grad_norm": 0.9424387812614441, + "learning_rate": 0.00019349009614270823, + "loss": 0.3166, + "step": 2530 + }, + { + "epoch": 0.09807328468280628, + "grad_norm": 3.792689800262451, + "learning_rate": 0.00019346435512310642, + "loss": 0.2591, + "step": 2540 + }, + { + "epoch": 0.09845939997683308, + "grad_norm": 2.9132003784179688, + "learning_rate": 0.00019343861410350466, + "loss": 0.3523, + "step": 2550 + }, + { + "epoch": 0.09884551527085988, + "grad_norm": 1.6421749591827393, + "learning_rate": 0.00019341287308390285, + "loss": 0.2805, + "step": 2560 + }, + { + "epoch": 0.09923163056488668, + "grad_norm": 1.9469813108444214, + "learning_rate": 0.0001933871320643011, + "loss": 0.3929, + "step": 2570 + }, + { + "epoch": 0.09961774585891348, + "grad_norm": 3.081062078475952, + "learning_rate": 0.00019336139104469927, + "loss": 0.569, + "step": 2580 + }, + { + "epoch": 0.10000386115294027, + "grad_norm": 4.728143215179443, + "learning_rate": 0.00019333565002509751, + "loss": 0.3443, + "step": 2590 + }, + { + "epoch": 0.10038997644696707, + "grad_norm": 2.4117422103881836, + "learning_rate": 0.00019330990900549573, + "loss": 0.4492, + "step": 2600 + }, + { + "epoch": 0.10077609174099386, + "grad_norm": 5.794168472290039, + "learning_rate": 0.00019328416798589391, + "loss": 0.5088, + "step": 2610 + }, + { + "epoch": 0.10116220703502066, + "grad_norm": 1.0662094354629517, + "learning_rate": 0.00019325842696629215, + "loss": 0.4683, + "step": 2620 + }, + { + "epoch": 0.10154832232904745, + "grad_norm": 2.13590931892395, + "learning_rate": 0.00019323268594669034, + "loss": 0.5833, + "step": 2630 + }, + { + "epoch": 0.10193443762307425, + "grad_norm": 7.52834415435791, + "learning_rate": 0.00019320694492708858, + "loss": 0.54, + "step": 2640 + }, + { + "epoch": 0.10232055291710104, + "grad_norm": 6.155847072601318, + "learning_rate": 0.00019318120390748677, + "loss": 0.6809, + "step": 2650 + }, + { + "epoch": 0.10270666821112784, + "grad_norm": 6.527890205383301, + "learning_rate": 0.000193155462887885, + "loss": 0.5239, + "step": 2660 + }, + { + "epoch": 0.10309278350515463, + "grad_norm": 3.3918986320495605, + "learning_rate": 0.00019312972186828322, + "loss": 0.4636, + "step": 2670 + }, + { + "epoch": 0.10347889879918143, + "grad_norm": 2.0933191776275635, + "learning_rate": 0.0001931039808486814, + "loss": 0.5288, + "step": 2680 + }, + { + "epoch": 0.10386501409320824, + "grad_norm": 2.4386465549468994, + "learning_rate": 0.00019307823982907965, + "loss": 0.5496, + "step": 2690 + }, + { + "epoch": 0.10425112938723503, + "grad_norm": 2.885315418243408, + "learning_rate": 0.00019305249880947783, + "loss": 0.5928, + "step": 2700 + }, + { + "epoch": 0.10463724468126183, + "grad_norm": 4.986598968505859, + "learning_rate": 0.00019302675778987607, + "loss": 0.3513, + "step": 2710 + }, + { + "epoch": 0.10502335997528862, + "grad_norm": 2.7999277114868164, + "learning_rate": 0.0001930010167702743, + "loss": 0.4484, + "step": 2720 + }, + { + "epoch": 0.10540947526931542, + "grad_norm": 1.6467676162719727, + "learning_rate": 0.0001929752757506725, + "loss": 0.4729, + "step": 2730 + }, + { + "epoch": 0.10579559056334222, + "grad_norm": 2.168477773666382, + "learning_rate": 0.00019294953473107071, + "loss": 0.3579, + "step": 2740 + }, + { + "epoch": 0.10618170585736901, + "grad_norm": 1.5439807176589966, + "learning_rate": 0.00019292379371146893, + "loss": 0.5715, + "step": 2750 + }, + { + "epoch": 0.1065678211513958, + "grad_norm": 1.2175947427749634, + "learning_rate": 0.00019289805269186714, + "loss": 0.4059, + "step": 2760 + }, + { + "epoch": 0.1069539364454226, + "grad_norm": 2.649782419204712, + "learning_rate": 0.00019287231167226533, + "loss": 0.7864, + "step": 2770 + }, + { + "epoch": 0.1073400517394494, + "grad_norm": 3.698002815246582, + "learning_rate": 0.00019284657065266357, + "loss": 0.5107, + "step": 2780 + }, + { + "epoch": 0.10772616703347619, + "grad_norm": 1.7418729066848755, + "learning_rate": 0.00019282082963306178, + "loss": 0.332, + "step": 2790 + }, + { + "epoch": 0.10811228232750299, + "grad_norm": 2.2264151573181152, + "learning_rate": 0.00019279508861346, + "loss": 0.3944, + "step": 2800 + }, + { + "epoch": 0.10849839762152978, + "grad_norm": 1.3407092094421387, + "learning_rate": 0.0001927693475938582, + "loss": 0.4603, + "step": 2810 + }, + { + "epoch": 0.10888451291555659, + "grad_norm": 0.3719189763069153, + "learning_rate": 0.00019274360657425642, + "loss": 0.496, + "step": 2820 + }, + { + "epoch": 0.10927062820958339, + "grad_norm": 4.728814125061035, + "learning_rate": 0.00019271786555465463, + "loss": 0.4079, + "step": 2830 + }, + { + "epoch": 0.10965674350361018, + "grad_norm": 5.007620334625244, + "learning_rate": 0.00019269212453505285, + "loss": 0.5221, + "step": 2840 + }, + { + "epoch": 0.11004285879763698, + "grad_norm": 2.7476320266723633, + "learning_rate": 0.00019266638351545106, + "loss": 0.3705, + "step": 2850 + }, + { + "epoch": 0.11042897409166377, + "grad_norm": 2.2960126399993896, + "learning_rate": 0.00019264064249584927, + "loss": 0.6656, + "step": 2860 + }, + { + "epoch": 0.11081508938569057, + "grad_norm": 0.9589812755584717, + "learning_rate": 0.0001926149014762475, + "loss": 0.6977, + "step": 2870 + }, + { + "epoch": 0.11120120467971736, + "grad_norm": 2.274984121322632, + "learning_rate": 0.0001925891604566457, + "loss": 0.4237, + "step": 2880 + }, + { + "epoch": 0.11158731997374416, + "grad_norm": 1.8849111795425415, + "learning_rate": 0.00019256341943704391, + "loss": 0.2357, + "step": 2890 + }, + { + "epoch": 0.11197343526777095, + "grad_norm": 2.7264740467071533, + "learning_rate": 0.00019253767841744213, + "loss": 0.3422, + "step": 2900 + }, + { + "epoch": 0.11235955056179775, + "grad_norm": 3.832827568054199, + "learning_rate": 0.00019251193739784034, + "loss": 0.3861, + "step": 2910 + }, + { + "epoch": 0.11274566585582455, + "grad_norm": 2.3612313270568848, + "learning_rate": 0.00019248619637823855, + "loss": 0.3299, + "step": 2920 + }, + { + "epoch": 0.11313178114985134, + "grad_norm": 2.2509396076202393, + "learning_rate": 0.00019246045535863677, + "loss": 0.6027, + "step": 2930 + }, + { + "epoch": 0.11351789644387814, + "grad_norm": 2.7600464820861816, + "learning_rate": 0.00019243471433903498, + "loss": 0.3864, + "step": 2940 + }, + { + "epoch": 0.11390401173790494, + "grad_norm": 5.956289768218994, + "learning_rate": 0.0001924089733194332, + "loss": 0.4669, + "step": 2950 + }, + { + "epoch": 0.11429012703193174, + "grad_norm": 4.651761531829834, + "learning_rate": 0.0001923832322998314, + "loss": 0.5774, + "step": 2960 + }, + { + "epoch": 0.11467624232595854, + "grad_norm": 1.1770590543746948, + "learning_rate": 0.00019235749128022962, + "loss": 0.3951, + "step": 2970 + }, + { + "epoch": 0.11506235761998533, + "grad_norm": 0.8117956519126892, + "learning_rate": 0.00019233175026062783, + "loss": 0.3919, + "step": 2980 + }, + { + "epoch": 0.11544847291401213, + "grad_norm": 1.318812608718872, + "learning_rate": 0.00019230600924102605, + "loss": 0.2905, + "step": 2990 + }, + { + "epoch": 0.11583458820803892, + "grad_norm": 1.981382966041565, + "learning_rate": 0.00019228026822142426, + "loss": 0.5312, + "step": 3000 + }, + { + "epoch": 0.11622070350206572, + "grad_norm": 1.9083003997802734, + "learning_rate": 0.00019225452720182247, + "loss": 0.3129, + "step": 3010 + }, + { + "epoch": 0.11660681879609251, + "grad_norm": 2.7581653594970703, + "learning_rate": 0.0001922287861822207, + "loss": 0.3954, + "step": 3020 + }, + { + "epoch": 0.11699293409011931, + "grad_norm": 1.065090537071228, + "learning_rate": 0.0001922030451626189, + "loss": 0.2722, + "step": 3030 + }, + { + "epoch": 0.1173790493841461, + "grad_norm": 0.600864827632904, + "learning_rate": 0.0001921773041430171, + "loss": 0.493, + "step": 3040 + }, + { + "epoch": 0.1177651646781729, + "grad_norm": 4.4449052810668945, + "learning_rate": 0.00019215156312341533, + "loss": 0.4982, + "step": 3050 + }, + { + "epoch": 0.1181512799721997, + "grad_norm": 3.34476637840271, + "learning_rate": 0.00019212582210381354, + "loss": 0.3204, + "step": 3060 + }, + { + "epoch": 0.11853739526622649, + "grad_norm": 1.7432445287704468, + "learning_rate": 0.00019210008108421175, + "loss": 0.6601, + "step": 3070 + }, + { + "epoch": 0.1189235105602533, + "grad_norm": 1.908324956893921, + "learning_rate": 0.00019207434006460997, + "loss": 0.5947, + "step": 3080 + }, + { + "epoch": 0.1193096258542801, + "grad_norm": 5.373056888580322, + "learning_rate": 0.00019204859904500818, + "loss": 0.5169, + "step": 3090 + }, + { + "epoch": 0.11969574114830689, + "grad_norm": 0.861535370349884, + "learning_rate": 0.0001920228580254064, + "loss": 0.3829, + "step": 3100 + }, + { + "epoch": 0.12008185644233368, + "grad_norm": 1.2700462341308594, + "learning_rate": 0.0001919971170058046, + "loss": 0.4475, + "step": 3110 + }, + { + "epoch": 0.12046797173636048, + "grad_norm": 2.9959444999694824, + "learning_rate": 0.00019197137598620282, + "loss": 0.4704, + "step": 3120 + }, + { + "epoch": 0.12085408703038727, + "grad_norm": 0.280109167098999, + "learning_rate": 0.00019194563496660103, + "loss": 0.3732, + "step": 3130 + }, + { + "epoch": 0.12124020232441407, + "grad_norm": 0.9746024012565613, + "learning_rate": 0.00019191989394699925, + "loss": 0.4693, + "step": 3140 + }, + { + "epoch": 0.12162631761844087, + "grad_norm": 1.7267721891403198, + "learning_rate": 0.00019189415292739746, + "loss": 0.4509, + "step": 3150 + }, + { + "epoch": 0.12201243291246766, + "grad_norm": 2.1759033203125, + "learning_rate": 0.00019186841190779567, + "loss": 0.428, + "step": 3160 + }, + { + "epoch": 0.12239854820649446, + "grad_norm": 1.270711064338684, + "learning_rate": 0.0001918426708881939, + "loss": 0.4262, + "step": 3170 + }, + { + "epoch": 0.12278466350052125, + "grad_norm": 3.7549123764038086, + "learning_rate": 0.0001918169298685921, + "loss": 0.4758, + "step": 3180 + }, + { + "epoch": 0.12317077879454805, + "grad_norm": 1.6550017595291138, + "learning_rate": 0.0001917911888489903, + "loss": 0.452, + "step": 3190 + }, + { + "epoch": 0.12355689408857484, + "grad_norm": 3.7151713371276855, + "learning_rate": 0.00019176544782938853, + "loss": 0.4844, + "step": 3200 + }, + { + "epoch": 0.12394300938260165, + "grad_norm": 0.5354440808296204, + "learning_rate": 0.00019173970680978674, + "loss": 0.4432, + "step": 3210 + }, + { + "epoch": 0.12432912467662845, + "grad_norm": 3.2494261264801025, + "learning_rate": 0.00019171396579018498, + "loss": 0.587, + "step": 3220 + }, + { + "epoch": 0.12471523997065524, + "grad_norm": 1.2129877805709839, + "learning_rate": 0.00019168822477058317, + "loss": 0.4662, + "step": 3230 + }, + { + "epoch": 0.12510135526468202, + "grad_norm": 3.723402500152588, + "learning_rate": 0.00019166248375098138, + "loss": 0.5261, + "step": 3240 + }, + { + "epoch": 0.12548747055870882, + "grad_norm": 1.596259593963623, + "learning_rate": 0.0001916367427313796, + "loss": 0.2802, + "step": 3250 + }, + { + "epoch": 0.12587358585273561, + "grad_norm": 5.5710320472717285, + "learning_rate": 0.0001916110017117778, + "loss": 0.5246, + "step": 3260 + }, + { + "epoch": 0.1262597011467624, + "grad_norm": 4.490183353424072, + "learning_rate": 0.00019158526069217602, + "loss": 0.4929, + "step": 3270 + }, + { + "epoch": 0.12664581644078923, + "grad_norm": 2.482572555541992, + "learning_rate": 0.00019155951967257423, + "loss": 0.3677, + "step": 3280 + }, + { + "epoch": 0.12703193173481603, + "grad_norm": 3.348520517349243, + "learning_rate": 0.00019153377865297247, + "loss": 0.6471, + "step": 3290 + }, + { + "epoch": 0.12741804702884282, + "grad_norm": 7.735306262969971, + "learning_rate": 0.00019150803763337066, + "loss": 0.6057, + "step": 3300 + }, + { + "epoch": 0.12780416232286962, + "grad_norm": 2.120649576187134, + "learning_rate": 0.00019148229661376887, + "loss": 0.5408, + "step": 3310 + }, + { + "epoch": 0.12819027761689641, + "grad_norm": 10.259540557861328, + "learning_rate": 0.00019145655559416709, + "loss": 0.4753, + "step": 3320 + }, + { + "epoch": 0.1285763929109232, + "grad_norm": 4.094576358795166, + "learning_rate": 0.0001914308145745653, + "loss": 0.3832, + "step": 3330 + }, + { + "epoch": 0.12896250820495, + "grad_norm": 3.1248559951782227, + "learning_rate": 0.00019140507355496354, + "loss": 0.5631, + "step": 3340 + }, + { + "epoch": 0.1293486234989768, + "grad_norm": 1.2975168228149414, + "learning_rate": 0.00019137933253536173, + "loss": 0.5158, + "step": 3350 + }, + { + "epoch": 0.1297347387930036, + "grad_norm": 3.2515244483947754, + "learning_rate": 0.00019135359151575997, + "loss": 0.4176, + "step": 3360 + }, + { + "epoch": 0.1301208540870304, + "grad_norm": 2.287757396697998, + "learning_rate": 0.00019132785049615815, + "loss": 0.5316, + "step": 3370 + }, + { + "epoch": 0.1305069693810572, + "grad_norm": 8.668967247009277, + "learning_rate": 0.00019130210947655637, + "loss": 0.6653, + "step": 3380 + }, + { + "epoch": 0.13089308467508398, + "grad_norm": 4.751536846160889, + "learning_rate": 0.00019127636845695458, + "loss": 0.4508, + "step": 3390 + }, + { + "epoch": 0.13127919996911078, + "grad_norm": 3.240792751312256, + "learning_rate": 0.0001912506274373528, + "loss": 0.441, + "step": 3400 + }, + { + "epoch": 0.13166531526313757, + "grad_norm": 2.146261215209961, + "learning_rate": 0.00019122488641775103, + "loss": 0.3394, + "step": 3410 + }, + { + "epoch": 0.13205143055716437, + "grad_norm": 2.259693145751953, + "learning_rate": 0.00019119914539814922, + "loss": 0.4348, + "step": 3420 + }, + { + "epoch": 0.13243754585119116, + "grad_norm": 1.8136098384857178, + "learning_rate": 0.00019117340437854746, + "loss": 0.4441, + "step": 3430 + }, + { + "epoch": 0.13282366114521796, + "grad_norm": 1.7324503660202026, + "learning_rate": 0.00019114766335894565, + "loss": 0.4725, + "step": 3440 + }, + { + "epoch": 0.13320977643924475, + "grad_norm": 4.709383487701416, + "learning_rate": 0.00019112192233934389, + "loss": 0.5383, + "step": 3450 + }, + { + "epoch": 0.13359589173327155, + "grad_norm": 0.3468118906021118, + "learning_rate": 0.00019109618131974207, + "loss": 0.3228, + "step": 3460 + }, + { + "epoch": 0.13398200702729834, + "grad_norm": 6.66448974609375, + "learning_rate": 0.00019107044030014029, + "loss": 0.4128, + "step": 3470 + }, + { + "epoch": 0.13436812232132514, + "grad_norm": 1.6971935033798218, + "learning_rate": 0.00019104469928053853, + "loss": 0.5061, + "step": 3480 + }, + { + "epoch": 0.13475423761535194, + "grad_norm": 0.9180198311805725, + "learning_rate": 0.0001910189582609367, + "loss": 0.4445, + "step": 3490 + }, + { + "epoch": 0.13514035290937873, + "grad_norm": 3.1244235038757324, + "learning_rate": 0.00019099321724133495, + "loss": 0.4099, + "step": 3500 + }, + { + "epoch": 0.13552646820340553, + "grad_norm": 2.7192864418029785, + "learning_rate": 0.00019096747622173314, + "loss": 0.5004, + "step": 3510 + }, + { + "epoch": 0.13591258349743232, + "grad_norm": 2.0905699729919434, + "learning_rate": 0.00019094173520213138, + "loss": 0.4277, + "step": 3520 + }, + { + "epoch": 0.13629869879145912, + "grad_norm": 3.3753092288970947, + "learning_rate": 0.0001909159941825296, + "loss": 0.4129, + "step": 3530 + }, + { + "epoch": 0.13668481408548594, + "grad_norm": 4.199211120605469, + "learning_rate": 0.00019089025316292778, + "loss": 0.5734, + "step": 3540 + }, + { + "epoch": 0.13707092937951273, + "grad_norm": 2.0411245822906494, + "learning_rate": 0.00019086451214332602, + "loss": 0.3018, + "step": 3550 + }, + { + "epoch": 0.13745704467353953, + "grad_norm": 17.236717224121094, + "learning_rate": 0.0001908387711237242, + "loss": 0.4527, + "step": 3560 + }, + { + "epoch": 0.13784315996756633, + "grad_norm": 1.4575644731521606, + "learning_rate": 0.00019081303010412245, + "loss": 0.3773, + "step": 3570 + }, + { + "epoch": 0.13822927526159312, + "grad_norm": 3.926090717315674, + "learning_rate": 0.00019078728908452063, + "loss": 0.5316, + "step": 3580 + }, + { + "epoch": 0.13861539055561992, + "grad_norm": 3.1841864585876465, + "learning_rate": 0.00019076154806491887, + "loss": 0.3705, + "step": 3590 + }, + { + "epoch": 0.1390015058496467, + "grad_norm": 4.08506441116333, + "learning_rate": 0.00019073580704531709, + "loss": 0.4941, + "step": 3600 + }, + { + "epoch": 0.1393876211436735, + "grad_norm": 3.063154458999634, + "learning_rate": 0.00019071006602571527, + "loss": 0.4435, + "step": 3610 + }, + { + "epoch": 0.1397737364377003, + "grad_norm": 6.122230529785156, + "learning_rate": 0.0001906843250061135, + "loss": 0.5067, + "step": 3620 + }, + { + "epoch": 0.1401598517317271, + "grad_norm": 3.3089540004730225, + "learning_rate": 0.0001906585839865117, + "loss": 0.4329, + "step": 3630 + }, + { + "epoch": 0.1405459670257539, + "grad_norm": 1.7245008945465088, + "learning_rate": 0.00019063284296690994, + "loss": 0.4502, + "step": 3640 + }, + { + "epoch": 0.1409320823197807, + "grad_norm": 1.7759568691253662, + "learning_rate": 0.00019060710194730813, + "loss": 0.2379, + "step": 3650 + }, + { + "epoch": 0.14131819761380748, + "grad_norm": 0.432452529668808, + "learning_rate": 0.00019058136092770637, + "loss": 0.4277, + "step": 3660 + }, + { + "epoch": 0.14170431290783428, + "grad_norm": 3.311952829360962, + "learning_rate": 0.00019055561990810458, + "loss": 0.4558, + "step": 3670 + }, + { + "epoch": 0.14209042820186107, + "grad_norm": 1.9942964315414429, + "learning_rate": 0.00019052987888850277, + "loss": 0.3349, + "step": 3680 + }, + { + "epoch": 0.14247654349588787, + "grad_norm": 6.226424217224121, + "learning_rate": 0.000190504137868901, + "loss": 0.5809, + "step": 3690 + }, + { + "epoch": 0.14286265878991466, + "grad_norm": 6.223634719848633, + "learning_rate": 0.0001904783968492992, + "loss": 0.5788, + "step": 3700 + }, + { + "epoch": 0.14324877408394146, + "grad_norm": 0.7370914220809937, + "learning_rate": 0.00019045265582969743, + "loss": 0.4834, + "step": 3710 + }, + { + "epoch": 0.14363488937796826, + "grad_norm": 1.745880365371704, + "learning_rate": 0.00019042691481009565, + "loss": 0.6995, + "step": 3720 + }, + { + "epoch": 0.14402100467199505, + "grad_norm": 0.8839595913887024, + "learning_rate": 0.00019040117379049386, + "loss": 0.3526, + "step": 3730 + }, + { + "epoch": 0.14440711996602185, + "grad_norm": 1.1224008798599243, + "learning_rate": 0.00019037543277089207, + "loss": 0.3558, + "step": 3740 + }, + { + "epoch": 0.14479323526004864, + "grad_norm": 1.0473041534423828, + "learning_rate": 0.00019034969175129026, + "loss": 0.2465, + "step": 3750 + }, + { + "epoch": 0.14517935055407544, + "grad_norm": 3.83192777633667, + "learning_rate": 0.0001903239507316885, + "loss": 0.4832, + "step": 3760 + }, + { + "epoch": 0.14556546584810223, + "grad_norm": 3.323885440826416, + "learning_rate": 0.00019029820971208669, + "loss": 0.4924, + "step": 3770 + }, + { + "epoch": 0.14595158114212903, + "grad_norm": 3.2334187030792236, + "learning_rate": 0.00019027246869248493, + "loss": 0.5053, + "step": 3780 + }, + { + "epoch": 0.14633769643615582, + "grad_norm": 2.280498743057251, + "learning_rate": 0.00019024672767288314, + "loss": 0.554, + "step": 3790 + }, + { + "epoch": 0.14672381173018265, + "grad_norm": 4.546648979187012, + "learning_rate": 0.00019022098665328135, + "loss": 0.3999, + "step": 3800 + }, + { + "epoch": 0.14710992702420944, + "grad_norm": 0.6303244829177856, + "learning_rate": 0.00019019524563367957, + "loss": 0.4481, + "step": 3810 + }, + { + "epoch": 0.14749604231823624, + "grad_norm": 2.605196475982666, + "learning_rate": 0.00019016950461407775, + "loss": 0.3561, + "step": 3820 + }, + { + "epoch": 0.14788215761226303, + "grad_norm": 3.0562639236450195, + "learning_rate": 0.000190143763594476, + "loss": 0.5903, + "step": 3830 + }, + { + "epoch": 0.14826827290628983, + "grad_norm": 11.164155006408691, + "learning_rate": 0.0001901180225748742, + "loss": 0.4299, + "step": 3840 + }, + { + "epoch": 0.14865438820031662, + "grad_norm": 4.996811866760254, + "learning_rate": 0.00019009228155527242, + "loss": 0.4423, + "step": 3850 + }, + { + "epoch": 0.14904050349434342, + "grad_norm": 2.627272844314575, + "learning_rate": 0.00019006654053567063, + "loss": 0.4875, + "step": 3860 + }, + { + "epoch": 0.1494266187883702, + "grad_norm": 2.6532809734344482, + "learning_rate": 0.00019004079951606885, + "loss": 0.5221, + "step": 3870 + }, + { + "epoch": 0.149812734082397, + "grad_norm": 5.821976661682129, + "learning_rate": 0.00019001505849646706, + "loss": 0.4793, + "step": 3880 + }, + { + "epoch": 0.1501988493764238, + "grad_norm": 2.888029098510742, + "learning_rate": 0.00018998931747686524, + "loss": 0.5784, + "step": 3890 + }, + { + "epoch": 0.1505849646704506, + "grad_norm": 0.9147624969482422, + "learning_rate": 0.00018996357645726349, + "loss": 0.5533, + "step": 3900 + }, + { + "epoch": 0.1509710799644774, + "grad_norm": 2.6088199615478516, + "learning_rate": 0.0001899378354376617, + "loss": 0.5028, + "step": 3910 + }, + { + "epoch": 0.1513571952585042, + "grad_norm": 3.8208296298980713, + "learning_rate": 0.0001899120944180599, + "loss": 0.4934, + "step": 3920 + }, + { + "epoch": 0.15174331055253099, + "grad_norm": 2.8711328506469727, + "learning_rate": 0.00018988635339845813, + "loss": 0.4417, + "step": 3930 + }, + { + "epoch": 0.15212942584655778, + "grad_norm": 2.922855854034424, + "learning_rate": 0.00018986061237885634, + "loss": 0.5303, + "step": 3940 + }, + { + "epoch": 0.15251554114058458, + "grad_norm": 2.52575945854187, + "learning_rate": 0.00018983487135925455, + "loss": 0.397, + "step": 3950 + }, + { + "epoch": 0.15290165643461137, + "grad_norm": 3.3369996547698975, + "learning_rate": 0.00018980913033965277, + "loss": 0.4172, + "step": 3960 + }, + { + "epoch": 0.15328777172863817, + "grad_norm": 1.7678214311599731, + "learning_rate": 0.00018978338932005098, + "loss": 0.3122, + "step": 3970 + }, + { + "epoch": 0.15367388702266496, + "grad_norm": 3.3293211460113525, + "learning_rate": 0.0001897576483004492, + "loss": 0.6864, + "step": 3980 + }, + { + "epoch": 0.15406000231669176, + "grad_norm": 1.4911530017852783, + "learning_rate": 0.0001897319072808474, + "loss": 0.3888, + "step": 3990 + }, + { + "epoch": 0.15444611761071855, + "grad_norm": 1.4884055852890015, + "learning_rate": 0.00018970616626124562, + "loss": 0.3952, + "step": 4000 + }, + { + "epoch": 0.15483223290474535, + "grad_norm": 1.2745383977890015, + "learning_rate": 0.00018968042524164383, + "loss": 0.3647, + "step": 4010 + }, + { + "epoch": 0.15521834819877214, + "grad_norm": 7.799386024475098, + "learning_rate": 0.00018965468422204205, + "loss": 0.5554, + "step": 4020 + }, + { + "epoch": 0.15560446349279894, + "grad_norm": 2.4778294563293457, + "learning_rate": 0.00018962894320244026, + "loss": 0.662, + "step": 4030 + }, + { + "epoch": 0.15599057878682573, + "grad_norm": 0.8415629267692566, + "learning_rate": 0.00018960320218283847, + "loss": 0.4317, + "step": 4040 + }, + { + "epoch": 0.15637669408085253, + "grad_norm": 4.507715702056885, + "learning_rate": 0.00018957746116323669, + "loss": 0.4512, + "step": 4050 + }, + { + "epoch": 0.15676280937487935, + "grad_norm": 3.5790421962738037, + "learning_rate": 0.0001895517201436349, + "loss": 0.4022, + "step": 4060 + }, + { + "epoch": 0.15714892466890615, + "grad_norm": 3.7266156673431396, + "learning_rate": 0.0001895259791240331, + "loss": 0.3945, + "step": 4070 + }, + { + "epoch": 0.15753503996293294, + "grad_norm": 7.909580230712891, + "learning_rate": 0.00018950023810443133, + "loss": 0.3726, + "step": 4080 + }, + { + "epoch": 0.15792115525695974, + "grad_norm": 2.2439534664154053, + "learning_rate": 0.00018947449708482954, + "loss": 0.4157, + "step": 4090 + }, + { + "epoch": 0.15830727055098653, + "grad_norm": 1.6076972484588623, + "learning_rate": 0.00018944875606522775, + "loss": 0.2363, + "step": 4100 + }, + { + "epoch": 0.15869338584501333, + "grad_norm": 3.7495157718658447, + "learning_rate": 0.00018942301504562596, + "loss": 0.4908, + "step": 4110 + }, + { + "epoch": 0.15907950113904012, + "grad_norm": 0.2942291796207428, + "learning_rate": 0.00018939727402602418, + "loss": 0.4915, + "step": 4120 + }, + { + "epoch": 0.15946561643306692, + "grad_norm": 1.3951829671859741, + "learning_rate": 0.0001893715330064224, + "loss": 0.4585, + "step": 4130 + }, + { + "epoch": 0.15985173172709372, + "grad_norm": 0.4405671954154968, + "learning_rate": 0.0001893457919868206, + "loss": 0.2839, + "step": 4140 + }, + { + "epoch": 0.1602378470211205, + "grad_norm": 1.0917588472366333, + "learning_rate": 0.00018932005096721882, + "loss": 0.401, + "step": 4150 + }, + { + "epoch": 0.1606239623151473, + "grad_norm": 1.6183397769927979, + "learning_rate": 0.00018929430994761703, + "loss": 0.5555, + "step": 4160 + }, + { + "epoch": 0.1610100776091741, + "grad_norm": 2.0909583568573, + "learning_rate": 0.00018926856892801524, + "loss": 0.52, + "step": 4170 + }, + { + "epoch": 0.1613961929032009, + "grad_norm": 2.901456356048584, + "learning_rate": 0.00018924282790841346, + "loss": 0.603, + "step": 4180 + }, + { + "epoch": 0.1617823081972277, + "grad_norm": 7.230431079864502, + "learning_rate": 0.00018921708688881167, + "loss": 0.6189, + "step": 4190 + }, + { + "epoch": 0.1621684234912545, + "grad_norm": 6.773900508880615, + "learning_rate": 0.00018919134586920988, + "loss": 0.2494, + "step": 4200 + }, + { + "epoch": 0.16255453878528128, + "grad_norm": 0.8557988405227661, + "learning_rate": 0.0001891656048496081, + "loss": 0.2617, + "step": 4210 + }, + { + "epoch": 0.16294065407930808, + "grad_norm": 1.3747268915176392, + "learning_rate": 0.00018913986383000634, + "loss": 0.4189, + "step": 4220 + }, + { + "epoch": 0.16332676937333487, + "grad_norm": 4.072261810302734, + "learning_rate": 0.00018911412281040452, + "loss": 0.5473, + "step": 4230 + }, + { + "epoch": 0.16371288466736167, + "grad_norm": 2.7210185527801514, + "learning_rate": 0.00018908838179080274, + "loss": 0.3501, + "step": 4240 + }, + { + "epoch": 0.16409899996138846, + "grad_norm": 2.276454448699951, + "learning_rate": 0.00018906264077120095, + "loss": 0.3078, + "step": 4250 + }, + { + "epoch": 0.16448511525541526, + "grad_norm": 3.586536169052124, + "learning_rate": 0.00018903689975159916, + "loss": 0.3856, + "step": 4260 + }, + { + "epoch": 0.16487123054944205, + "grad_norm": 2.199673891067505, + "learning_rate": 0.00018901115873199738, + "loss": 0.3677, + "step": 4270 + }, + { + "epoch": 0.16525734584346885, + "grad_norm": 2.8410561084747314, + "learning_rate": 0.0001889854177123956, + "loss": 0.6101, + "step": 4280 + }, + { + "epoch": 0.16564346113749565, + "grad_norm": 3.9638853073120117, + "learning_rate": 0.00018895967669279383, + "loss": 0.5066, + "step": 4290 + }, + { + "epoch": 0.16602957643152244, + "grad_norm": 1.2070738077163696, + "learning_rate": 0.00018893393567319202, + "loss": 0.385, + "step": 4300 + }, + { + "epoch": 0.16641569172554924, + "grad_norm": 1.0531187057495117, + "learning_rate": 0.00018890819465359023, + "loss": 0.3608, + "step": 4310 + }, + { + "epoch": 0.16680180701957603, + "grad_norm": 1.1998246908187866, + "learning_rate": 0.00018888245363398844, + "loss": 0.4624, + "step": 4320 + }, + { + "epoch": 0.16718792231360285, + "grad_norm": 2.126063346862793, + "learning_rate": 0.00018885671261438666, + "loss": 0.6076, + "step": 4330 + }, + { + "epoch": 0.16757403760762965, + "grad_norm": 1.5854765176773071, + "learning_rate": 0.0001888309715947849, + "loss": 0.4817, + "step": 4340 + }, + { + "epoch": 0.16796015290165645, + "grad_norm": 6.630712509155273, + "learning_rate": 0.00018880523057518308, + "loss": 0.4098, + "step": 4350 + }, + { + "epoch": 0.16834626819568324, + "grad_norm": 2.060789108276367, + "learning_rate": 0.00018877948955558132, + "loss": 0.3523, + "step": 4360 + }, + { + "epoch": 0.16873238348971004, + "grad_norm": 2.2551252841949463, + "learning_rate": 0.0001887537485359795, + "loss": 0.3095, + "step": 4370 + }, + { + "epoch": 0.16911849878373683, + "grad_norm": 3.736640453338623, + "learning_rate": 0.00018872800751637772, + "loss": 0.3812, + "step": 4380 + }, + { + "epoch": 0.16950461407776363, + "grad_norm": 1.9971100091934204, + "learning_rate": 0.00018870226649677594, + "loss": 0.3422, + "step": 4390 + }, + { + "epoch": 0.16989072937179042, + "grad_norm": 3.6577255725860596, + "learning_rate": 0.00018867652547717415, + "loss": 0.7857, + "step": 4400 + }, + { + "epoch": 0.17027684466581722, + "grad_norm": 2.166538715362549, + "learning_rate": 0.0001886507844575724, + "loss": 0.5596, + "step": 4410 + }, + { + "epoch": 0.170662959959844, + "grad_norm": 2.0177736282348633, + "learning_rate": 0.00018862504343797058, + "loss": 0.3197, + "step": 4420 + }, + { + "epoch": 0.1710490752538708, + "grad_norm": 0.29447808861732483, + "learning_rate": 0.00018859930241836882, + "loss": 0.5284, + "step": 4430 + }, + { + "epoch": 0.1714351905478976, + "grad_norm": 2.17985200881958, + "learning_rate": 0.000188573561398767, + "loss": 0.5188, + "step": 4440 + }, + { + "epoch": 0.1718213058419244, + "grad_norm": 2.87449049949646, + "learning_rate": 0.00018854782037916522, + "loss": 0.554, + "step": 4450 + }, + { + "epoch": 0.1722074211359512, + "grad_norm": 1.8865265846252441, + "learning_rate": 0.00018852207935956343, + "loss": 0.4338, + "step": 4460 + }, + { + "epoch": 0.172593536429978, + "grad_norm": 2.042337417602539, + "learning_rate": 0.00018849633833996164, + "loss": 0.3924, + "step": 4470 + }, + { + "epoch": 0.17297965172400478, + "grad_norm": 1.4254354238510132, + "learning_rate": 0.00018847059732035988, + "loss": 0.2607, + "step": 4480 + }, + { + "epoch": 0.17336576701803158, + "grad_norm": 2.611560344696045, + "learning_rate": 0.00018844485630075807, + "loss": 0.4967, + "step": 4490 + }, + { + "epoch": 0.17375188231205838, + "grad_norm": 1.1008936166763306, + "learning_rate": 0.0001884191152811563, + "loss": 0.4109, + "step": 4500 + }, + { + "epoch": 0.17413799760608517, + "grad_norm": 0.8280178308486938, + "learning_rate": 0.0001883933742615545, + "loss": 0.6632, + "step": 4510 + }, + { + "epoch": 0.17452411290011197, + "grad_norm": 2.226020336151123, + "learning_rate": 0.0001883676332419527, + "loss": 0.4777, + "step": 4520 + }, + { + "epoch": 0.17491022819413876, + "grad_norm": 1.6062042713165283, + "learning_rate": 0.00018834189222235095, + "loss": 0.4671, + "step": 4530 + }, + { + "epoch": 0.17529634348816556, + "grad_norm": 3.9853012561798096, + "learning_rate": 0.00018831615120274914, + "loss": 0.4843, + "step": 4540 + }, + { + "epoch": 0.17568245878219235, + "grad_norm": 0.30268657207489014, + "learning_rate": 0.00018829041018314738, + "loss": 0.3922, + "step": 4550 + }, + { + "epoch": 0.17606857407621915, + "grad_norm": 6.283960342407227, + "learning_rate": 0.00018826466916354556, + "loss": 0.6106, + "step": 4560 + }, + { + "epoch": 0.17645468937024594, + "grad_norm": 1.4164658784866333, + "learning_rate": 0.0001882389281439438, + "loss": 0.3014, + "step": 4570 + }, + { + "epoch": 0.17684080466427274, + "grad_norm": 4.847668170928955, + "learning_rate": 0.000188213187124342, + "loss": 0.5216, + "step": 4580 + }, + { + "epoch": 0.17722691995829956, + "grad_norm": 3.683180332183838, + "learning_rate": 0.0001881874461047402, + "loss": 0.3268, + "step": 4590 + }, + { + "epoch": 0.17761303525232636, + "grad_norm": 1.053144097328186, + "learning_rate": 0.00018816170508513844, + "loss": 0.5229, + "step": 4600 + }, + { + "epoch": 0.17799915054635315, + "grad_norm": 0.29438719153404236, + "learning_rate": 0.00018813596406553663, + "loss": 0.4523, + "step": 4610 + }, + { + "epoch": 0.17838526584037995, + "grad_norm": 1.5682024955749512, + "learning_rate": 0.00018811022304593487, + "loss": 0.4367, + "step": 4620 + }, + { + "epoch": 0.17877138113440674, + "grad_norm": 1.462189793586731, + "learning_rate": 0.00018808448202633306, + "loss": 0.5086, + "step": 4630 + }, + { + "epoch": 0.17915749642843354, + "grad_norm": 0.7927210927009583, + "learning_rate": 0.0001880587410067313, + "loss": 0.4654, + "step": 4640 + }, + { + "epoch": 0.17954361172246033, + "grad_norm": 1.4543548822402954, + "learning_rate": 0.0001880329999871295, + "loss": 0.5005, + "step": 4650 + }, + { + "epoch": 0.17992972701648713, + "grad_norm": 1.5814868211746216, + "learning_rate": 0.00018800725896752772, + "loss": 0.4127, + "step": 4660 + }, + { + "epoch": 0.18031584231051392, + "grad_norm": 1.9244798421859741, + "learning_rate": 0.00018798151794792594, + "loss": 0.3796, + "step": 4670 + }, + { + "epoch": 0.18070195760454072, + "grad_norm": 1.8725996017456055, + "learning_rate": 0.00018795577692832412, + "loss": 0.4112, + "step": 4680 + }, + { + "epoch": 0.18108807289856751, + "grad_norm": 2.8138442039489746, + "learning_rate": 0.00018793003590872236, + "loss": 0.6117, + "step": 4690 + }, + { + "epoch": 0.1814741881925943, + "grad_norm": 3.4465060234069824, + "learning_rate": 0.00018790429488912055, + "loss": 0.4223, + "step": 4700 + }, + { + "epoch": 0.1818603034866211, + "grad_norm": 4.431785583496094, + "learning_rate": 0.0001878785538695188, + "loss": 0.54, + "step": 4710 + }, + { + "epoch": 0.1822464187806479, + "grad_norm": 6.951846599578857, + "learning_rate": 0.000187852812849917, + "loss": 0.3702, + "step": 4720 + }, + { + "epoch": 0.1826325340746747, + "grad_norm": 1.0188024044036865, + "learning_rate": 0.00018782707183031522, + "loss": 0.2715, + "step": 4730 + }, + { + "epoch": 0.1830186493687015, + "grad_norm": 0.3875834047794342, + "learning_rate": 0.00018780133081071343, + "loss": 0.4208, + "step": 4740 + }, + { + "epoch": 0.1834047646627283, + "grad_norm": 2.7475740909576416, + "learning_rate": 0.00018777558979111162, + "loss": 0.3613, + "step": 4750 + }, + { + "epoch": 0.18379087995675508, + "grad_norm": 2.553227186203003, + "learning_rate": 0.00018774984877150986, + "loss": 0.4781, + "step": 4760 + }, + { + "epoch": 0.18417699525078188, + "grad_norm": 2.005154609680176, + "learning_rate": 0.00018772410775190804, + "loss": 0.3805, + "step": 4770 + }, + { + "epoch": 0.18456311054480867, + "grad_norm": 0.7380127310752869, + "learning_rate": 0.00018769836673230628, + "loss": 0.3679, + "step": 4780 + }, + { + "epoch": 0.18494922583883547, + "grad_norm": 3.6547505855560303, + "learning_rate": 0.0001876726257127045, + "loss": 0.4502, + "step": 4790 + }, + { + "epoch": 0.18533534113286226, + "grad_norm": 2.232980728149414, + "learning_rate": 0.0001876468846931027, + "loss": 0.4628, + "step": 4800 + }, + { + "epoch": 0.18572145642688906, + "grad_norm": 6.521275043487549, + "learning_rate": 0.00018762114367350092, + "loss": 0.4765, + "step": 4810 + }, + { + "epoch": 0.18610757172091585, + "grad_norm": 1.6310979127883911, + "learning_rate": 0.0001875954026538991, + "loss": 0.4039, + "step": 4820 + }, + { + "epoch": 0.18649368701494265, + "grad_norm": 1.1469775438308716, + "learning_rate": 0.00018756966163429735, + "loss": 0.4195, + "step": 4830 + }, + { + "epoch": 0.18687980230896944, + "grad_norm": 0.7688332200050354, + "learning_rate": 0.00018754392061469556, + "loss": 0.264, + "step": 4840 + }, + { + "epoch": 0.18726591760299627, + "grad_norm": 3.3422155380249023, + "learning_rate": 0.00018751817959509378, + "loss": 0.5275, + "step": 4850 + }, + { + "epoch": 0.18765203289702306, + "grad_norm": 1.517876386642456, + "learning_rate": 0.000187492438575492, + "loss": 0.4567, + "step": 4860 + }, + { + "epoch": 0.18803814819104986, + "grad_norm": 1.2196050882339478, + "learning_rate": 0.0001874666975558902, + "loss": 0.4231, + "step": 4870 + }, + { + "epoch": 0.18842426348507665, + "grad_norm": 1.3325402736663818, + "learning_rate": 0.00018744095653628842, + "loss": 0.6325, + "step": 4880 + }, + { + "epoch": 0.18881037877910345, + "grad_norm": 6.098769664764404, + "learning_rate": 0.0001874152155166866, + "loss": 0.576, + "step": 4890 + }, + { + "epoch": 0.18919649407313024, + "grad_norm": 2.602363348007202, + "learning_rate": 0.00018738947449708484, + "loss": 0.3237, + "step": 4900 + }, + { + "epoch": 0.18958260936715704, + "grad_norm": 0.970106303691864, + "learning_rate": 0.00018736373347748306, + "loss": 0.409, + "step": 4910 + }, + { + "epoch": 0.18996872466118384, + "grad_norm": 3.2592012882232666, + "learning_rate": 0.00018733799245788127, + "loss": 0.408, + "step": 4920 + }, + { + "epoch": 0.19035483995521063, + "grad_norm": 0.31132128834724426, + "learning_rate": 0.00018731225143827948, + "loss": 0.2446, + "step": 4930 + }, + { + "epoch": 0.19074095524923743, + "grad_norm": 5.321741104125977, + "learning_rate": 0.0001872865104186777, + "loss": 0.4604, + "step": 4940 + }, + { + "epoch": 0.19112707054326422, + "grad_norm": 1.1165122985839844, + "learning_rate": 0.0001872607693990759, + "loss": 0.3605, + "step": 4950 + }, + { + "epoch": 0.19151318583729102, + "grad_norm": 0.8274110555648804, + "learning_rate": 0.0001872350283794741, + "loss": 0.2669, + "step": 4960 + }, + { + "epoch": 0.1918993011313178, + "grad_norm": 2.8668346405029297, + "learning_rate": 0.00018720928735987234, + "loss": 0.4055, + "step": 4970 + }, + { + "epoch": 0.1922854164253446, + "grad_norm": 3.411841630935669, + "learning_rate": 0.00018718354634027055, + "loss": 0.5989, + "step": 4980 + }, + { + "epoch": 0.1926715317193714, + "grad_norm": 0.18740829825401306, + "learning_rate": 0.00018715780532066876, + "loss": 0.3805, + "step": 4990 + }, + { + "epoch": 0.1930576470133982, + "grad_norm": 1.0823473930358887, + "learning_rate": 0.00018713206430106698, + "loss": 0.2854, + "step": 5000 + }, + { + "epoch": 0.193443762307425, + "grad_norm": 1.9816405773162842, + "learning_rate": 0.0001871063232814652, + "loss": 0.3771, + "step": 5010 + }, + { + "epoch": 0.1938298776014518, + "grad_norm": 5.267081260681152, + "learning_rate": 0.0001870805822618634, + "loss": 0.3085, + "step": 5020 + }, + { + "epoch": 0.19421599289547858, + "grad_norm": 5.706038475036621, + "learning_rate": 0.00018705484124226162, + "loss": 0.484, + "step": 5030 + }, + { + "epoch": 0.19460210818950538, + "grad_norm": 1.3357723951339722, + "learning_rate": 0.00018702910022265983, + "loss": 0.2161, + "step": 5040 + }, + { + "epoch": 0.19498822348353217, + "grad_norm": 1.0626447200775146, + "learning_rate": 0.00018700335920305804, + "loss": 0.3491, + "step": 5050 + }, + { + "epoch": 0.19537433877755897, + "grad_norm": 2.441228151321411, + "learning_rate": 0.00018697761818345626, + "loss": 0.3975, + "step": 5060 + }, + { + "epoch": 0.19576045407158577, + "grad_norm": 2.6739327907562256, + "learning_rate": 0.00018695187716385447, + "loss": 0.3418, + "step": 5070 + }, + { + "epoch": 0.19614656936561256, + "grad_norm": 2.3216919898986816, + "learning_rate": 0.00018692613614425268, + "loss": 0.6265, + "step": 5080 + }, + { + "epoch": 0.19653268465963936, + "grad_norm": 3.9119021892547607, + "learning_rate": 0.0001869003951246509, + "loss": 0.2982, + "step": 5090 + }, + { + "epoch": 0.19691879995366615, + "grad_norm": 5.744061470031738, + "learning_rate": 0.0001868746541050491, + "loss": 0.4048, + "step": 5100 + }, + { + "epoch": 0.19730491524769297, + "grad_norm": 8.512910842895508, + "learning_rate": 0.00018684891308544732, + "loss": 0.3598, + "step": 5110 + }, + { + "epoch": 0.19769103054171977, + "grad_norm": 1.6382296085357666, + "learning_rate": 0.00018682317206584554, + "loss": 0.2121, + "step": 5120 + }, + { + "epoch": 0.19807714583574657, + "grad_norm": 2.1593070030212402, + "learning_rate": 0.00018679743104624375, + "loss": 0.4914, + "step": 5130 + }, + { + "epoch": 0.19846326112977336, + "grad_norm": 3.067112445831299, + "learning_rate": 0.00018677169002664196, + "loss": 0.4171, + "step": 5140 + }, + { + "epoch": 0.19884937642380016, + "grad_norm": 1.9954415559768677, + "learning_rate": 0.00018674594900704018, + "loss": 0.5161, + "step": 5150 + }, + { + "epoch": 0.19923549171782695, + "grad_norm": 2.793346643447876, + "learning_rate": 0.0001867202079874384, + "loss": 0.4159, + "step": 5160 + }, + { + "epoch": 0.19962160701185375, + "grad_norm": 4.087403774261475, + "learning_rate": 0.0001866944669678366, + "loss": 0.3339, + "step": 5170 + }, + { + "epoch": 0.20000772230588054, + "grad_norm": 2.10153865814209, + "learning_rate": 0.00018666872594823482, + "loss": 0.4352, + "step": 5180 + }, + { + "epoch": 0.20039383759990734, + "grad_norm": 2.947117805480957, + "learning_rate": 0.00018664298492863303, + "loss": 0.305, + "step": 5190 + }, + { + "epoch": 0.20077995289393413, + "grad_norm": 1.2496302127838135, + "learning_rate": 0.00018661724390903124, + "loss": 0.4578, + "step": 5200 + }, + { + "epoch": 0.20116606818796093, + "grad_norm": 0.5246118903160095, + "learning_rate": 0.00018659150288942946, + "loss": 0.7531, + "step": 5210 + }, + { + "epoch": 0.20155218348198772, + "grad_norm": 4.099668502807617, + "learning_rate": 0.00018656576186982767, + "loss": 0.3809, + "step": 5220 + }, + { + "epoch": 0.20193829877601452, + "grad_norm": 4.237419128417969, + "learning_rate": 0.00018654002085022588, + "loss": 0.3169, + "step": 5230 + }, + { + "epoch": 0.20232441407004131, + "grad_norm": 1.6228466033935547, + "learning_rate": 0.0001865142798306241, + "loss": 0.5832, + "step": 5240 + }, + { + "epoch": 0.2027105293640681, + "grad_norm": 4.567386627197266, + "learning_rate": 0.0001864885388110223, + "loss": 0.2177, + "step": 5250 + }, + { + "epoch": 0.2030966446580949, + "grad_norm": 1.4991040229797363, + "learning_rate": 0.00018646279779142052, + "loss": 0.3851, + "step": 5260 + }, + { + "epoch": 0.2034827599521217, + "grad_norm": 2.127082586288452, + "learning_rate": 0.00018643705677181874, + "loss": 0.4721, + "step": 5270 + }, + { + "epoch": 0.2038688752461485, + "grad_norm": 2.9149303436279297, + "learning_rate": 0.00018641131575221695, + "loss": 0.2556, + "step": 5280 + }, + { + "epoch": 0.2042549905401753, + "grad_norm": 0.06375914812088013, + "learning_rate": 0.00018638557473261516, + "loss": 0.3599, + "step": 5290 + }, + { + "epoch": 0.20464110583420209, + "grad_norm": 3.338331699371338, + "learning_rate": 0.00018635983371301338, + "loss": 0.4062, + "step": 5300 + }, + { + "epoch": 0.20502722112822888, + "grad_norm": 4.006681442260742, + "learning_rate": 0.0001863340926934116, + "loss": 0.4538, + "step": 5310 + }, + { + "epoch": 0.20541333642225568, + "grad_norm": 1.1406009197235107, + "learning_rate": 0.0001863083516738098, + "loss": 0.8432, + "step": 5320 + }, + { + "epoch": 0.20579945171628247, + "grad_norm": 9.281437873840332, + "learning_rate": 0.00018628261065420802, + "loss": 0.4538, + "step": 5330 + }, + { + "epoch": 0.20618556701030927, + "grad_norm": 3.1884214878082275, + "learning_rate": 0.00018625686963460626, + "loss": 0.3361, + "step": 5340 + }, + { + "epoch": 0.20657168230433606, + "grad_norm": 1.4311977624893188, + "learning_rate": 0.00018623112861500444, + "loss": 0.5519, + "step": 5350 + }, + { + "epoch": 0.20695779759836286, + "grad_norm": 3.574361801147461, + "learning_rate": 0.00018620538759540266, + "loss": 0.518, + "step": 5360 + }, + { + "epoch": 0.20734391289238968, + "grad_norm": 3.0186073780059814, + "learning_rate": 0.00018617964657580087, + "loss": 0.4204, + "step": 5370 + }, + { + "epoch": 0.20773002818641648, + "grad_norm": 2.832859754562378, + "learning_rate": 0.00018615390555619908, + "loss": 0.5736, + "step": 5380 + }, + { + "epoch": 0.20811614348044327, + "grad_norm": 2.2258200645446777, + "learning_rate": 0.0001861281645365973, + "loss": 0.8194, + "step": 5390 + }, + { + "epoch": 0.20850225877447007, + "grad_norm": 1.0975148677825928, + "learning_rate": 0.0001861024235169955, + "loss": 0.5235, + "step": 5400 + }, + { + "epoch": 0.20888837406849686, + "grad_norm": 2.597329616546631, + "learning_rate": 0.00018607668249739375, + "loss": 0.2798, + "step": 5410 + }, + { + "epoch": 0.20927448936252366, + "grad_norm": 1.3780876398086548, + "learning_rate": 0.00018605094147779194, + "loss": 0.4046, + "step": 5420 + }, + { + "epoch": 0.20966060465655045, + "grad_norm": 2.409886598587036, + "learning_rate": 0.00018602520045819018, + "loss": 0.3243, + "step": 5430 + }, + { + "epoch": 0.21004671995057725, + "grad_norm": 1.0368077754974365, + "learning_rate": 0.00018599945943858836, + "loss": 0.4469, + "step": 5440 + }, + { + "epoch": 0.21043283524460404, + "grad_norm": 2.961658000946045, + "learning_rate": 0.00018597371841898658, + "loss": 0.5104, + "step": 5450 + }, + { + "epoch": 0.21081895053863084, + "grad_norm": 1.1599836349487305, + "learning_rate": 0.00018594797739938482, + "loss": 0.3422, + "step": 5460 + }, + { + "epoch": 0.21120506583265763, + "grad_norm": 3.293682336807251, + "learning_rate": 0.000185922236379783, + "loss": 0.3556, + "step": 5470 + }, + { + "epoch": 0.21159118112668443, + "grad_norm": 1.6923863887786865, + "learning_rate": 0.00018589649536018124, + "loss": 0.3084, + "step": 5480 + }, + { + "epoch": 0.21197729642071123, + "grad_norm": 3.7289531230926514, + "learning_rate": 0.00018587075434057943, + "loss": 0.4668, + "step": 5490 + }, + { + "epoch": 0.21236341171473802, + "grad_norm": 1.3744993209838867, + "learning_rate": 0.00018584501332097767, + "loss": 0.2984, + "step": 5500 + }, + { + "epoch": 0.21274952700876482, + "grad_norm": 1.4377775192260742, + "learning_rate": 0.00018581927230137586, + "loss": 0.2622, + "step": 5510 + }, + { + "epoch": 0.2131356423027916, + "grad_norm": 4.957859992980957, + "learning_rate": 0.00018579353128177407, + "loss": 0.5561, + "step": 5520 + }, + { + "epoch": 0.2135217575968184, + "grad_norm": 3.2645647525787354, + "learning_rate": 0.0001857677902621723, + "loss": 0.59, + "step": 5530 + }, + { + "epoch": 0.2139078728908452, + "grad_norm": 1.1365091800689697, + "learning_rate": 0.0001857420492425705, + "loss": 0.443, + "step": 5540 + }, + { + "epoch": 0.214293988184872, + "grad_norm": 3.187476396560669, + "learning_rate": 0.00018571630822296874, + "loss": 0.2612, + "step": 5550 + }, + { + "epoch": 0.2146801034788988, + "grad_norm": 2.6851940155029297, + "learning_rate": 0.00018569056720336692, + "loss": 0.4543, + "step": 5560 + }, + { + "epoch": 0.2150662187729256, + "grad_norm": 2.2613587379455566, + "learning_rate": 0.00018566482618376516, + "loss": 0.3185, + "step": 5570 + }, + { + "epoch": 0.21545233406695238, + "grad_norm": 1.292475700378418, + "learning_rate": 0.00018563908516416335, + "loss": 0.2794, + "step": 5580 + }, + { + "epoch": 0.21583844936097918, + "grad_norm": 2.0878446102142334, + "learning_rate": 0.00018561334414456156, + "loss": 0.3908, + "step": 5590 + }, + { + "epoch": 0.21622456465500597, + "grad_norm": 8.058819770812988, + "learning_rate": 0.0001855876031249598, + "loss": 0.6282, + "step": 5600 + }, + { + "epoch": 0.21661067994903277, + "grad_norm": 1.8231629133224487, + "learning_rate": 0.000185561862105358, + "loss": 0.4973, + "step": 5610 + }, + { + "epoch": 0.21699679524305956, + "grad_norm": 3.947242259979248, + "learning_rate": 0.00018553612108575623, + "loss": 0.4598, + "step": 5620 + }, + { + "epoch": 0.21738291053708636, + "grad_norm": 3.3258073329925537, + "learning_rate": 0.00018551038006615442, + "loss": 0.5266, + "step": 5630 + }, + { + "epoch": 0.21776902583111318, + "grad_norm": 2.301485300064087, + "learning_rate": 0.00018548463904655266, + "loss": 0.4339, + "step": 5640 + }, + { + "epoch": 0.21815514112513998, + "grad_norm": 4.4706878662109375, + "learning_rate": 0.00018545889802695087, + "loss": 0.5233, + "step": 5650 + }, + { + "epoch": 0.21854125641916677, + "grad_norm": 1.1203399896621704, + "learning_rate": 0.00018543315700734906, + "loss": 0.4547, + "step": 5660 + }, + { + "epoch": 0.21892737171319357, + "grad_norm": 0.3744584918022156, + "learning_rate": 0.0001854074159877473, + "loss": 0.2524, + "step": 5670 + }, + { + "epoch": 0.21931348700722036, + "grad_norm": 2.7888870239257812, + "learning_rate": 0.00018538167496814548, + "loss": 0.411, + "step": 5680 + }, + { + "epoch": 0.21969960230124716, + "grad_norm": 4.9972429275512695, + "learning_rate": 0.00018535593394854372, + "loss": 0.6359, + "step": 5690 + }, + { + "epoch": 0.22008571759527396, + "grad_norm": 1.1321420669555664, + "learning_rate": 0.0001853301929289419, + "loss": 0.4068, + "step": 5700 + }, + { + "epoch": 0.22047183288930075, + "grad_norm": 1.9291785955429077, + "learning_rate": 0.00018530445190934015, + "loss": 0.5428, + "step": 5710 + }, + { + "epoch": 0.22085794818332755, + "grad_norm": 0.8663263916969299, + "learning_rate": 0.00018527871088973836, + "loss": 0.4662, + "step": 5720 + }, + { + "epoch": 0.22124406347735434, + "grad_norm": 3.039782762527466, + "learning_rate": 0.00018525296987013655, + "loss": 0.3045, + "step": 5730 + }, + { + "epoch": 0.22163017877138114, + "grad_norm": 1.3552179336547852, + "learning_rate": 0.0001852272288505348, + "loss": 0.3411, + "step": 5740 + }, + { + "epoch": 0.22201629406540793, + "grad_norm": 1.4136948585510254, + "learning_rate": 0.00018520148783093298, + "loss": 0.5517, + "step": 5750 + }, + { + "epoch": 0.22240240935943473, + "grad_norm": 2.463942766189575, + "learning_rate": 0.00018517574681133122, + "loss": 0.4681, + "step": 5760 + }, + { + "epoch": 0.22278852465346152, + "grad_norm": 0.9063917994499207, + "learning_rate": 0.0001851500057917294, + "loss": 0.4537, + "step": 5770 + }, + { + "epoch": 0.22317463994748832, + "grad_norm": 2.352678060531616, + "learning_rate": 0.00018512426477212764, + "loss": 0.4245, + "step": 5780 + }, + { + "epoch": 0.2235607552415151, + "grad_norm": 2.0424869060516357, + "learning_rate": 0.00018509852375252586, + "loss": 0.2892, + "step": 5790 + }, + { + "epoch": 0.2239468705355419, + "grad_norm": 2.7604904174804688, + "learning_rate": 0.00018507278273292404, + "loss": 0.3606, + "step": 5800 + }, + { + "epoch": 0.2243329858295687, + "grad_norm": 2.827798366546631, + "learning_rate": 0.00018504704171332228, + "loss": 0.3212, + "step": 5810 + }, + { + "epoch": 0.2247191011235955, + "grad_norm": 3.1988680362701416, + "learning_rate": 0.00018502130069372047, + "loss": 0.5649, + "step": 5820 + }, + { + "epoch": 0.2251052164176223, + "grad_norm": 1.8216092586517334, + "learning_rate": 0.0001849955596741187, + "loss": 0.2871, + "step": 5830 + }, + { + "epoch": 0.2254913317116491, + "grad_norm": 2.7595627307891846, + "learning_rate": 0.00018496981865451692, + "loss": 0.665, + "step": 5840 + }, + { + "epoch": 0.22587744700567589, + "grad_norm": 1.2395098209381104, + "learning_rate": 0.00018494407763491514, + "loss": 0.2504, + "step": 5850 + }, + { + "epoch": 0.22626356229970268, + "grad_norm": 0.6991098523139954, + "learning_rate": 0.00018491833661531335, + "loss": 0.2263, + "step": 5860 + }, + { + "epoch": 0.22664967759372948, + "grad_norm": 11.053647994995117, + "learning_rate": 0.00018489259559571156, + "loss": 0.5919, + "step": 5870 + }, + { + "epoch": 0.22703579288775627, + "grad_norm": 2.8663880825042725, + "learning_rate": 0.00018486685457610978, + "loss": 0.3399, + "step": 5880 + }, + { + "epoch": 0.22742190818178307, + "grad_norm": 1.4995262622833252, + "learning_rate": 0.00018484111355650796, + "loss": 0.4474, + "step": 5890 + }, + { + "epoch": 0.2278080234758099, + "grad_norm": 3.275681972503662, + "learning_rate": 0.0001848153725369062, + "loss": 0.4347, + "step": 5900 + }, + { + "epoch": 0.22819413876983669, + "grad_norm": 14.772253036499023, + "learning_rate": 0.00018478963151730442, + "loss": 0.3705, + "step": 5910 + }, + { + "epoch": 0.22858025406386348, + "grad_norm": 3.184976816177368, + "learning_rate": 0.00018476389049770263, + "loss": 0.3866, + "step": 5920 + }, + { + "epoch": 0.22896636935789028, + "grad_norm": 2.310765504837036, + "learning_rate": 0.00018473814947810084, + "loss": 0.2717, + "step": 5930 + }, + { + "epoch": 0.22935248465191707, + "grad_norm": 2.061189889907837, + "learning_rate": 0.00018471240845849906, + "loss": 0.2054, + "step": 5940 + }, + { + "epoch": 0.22973859994594387, + "grad_norm": 10.815469741821289, + "learning_rate": 0.00018468666743889727, + "loss": 0.5868, + "step": 5950 + }, + { + "epoch": 0.23012471523997066, + "grad_norm": 1.7080497741699219, + "learning_rate": 0.00018466092641929548, + "loss": 0.236, + "step": 5960 + }, + { + "epoch": 0.23051083053399746, + "grad_norm": 7.389080047607422, + "learning_rate": 0.0001846351853996937, + "loss": 0.2752, + "step": 5970 + }, + { + "epoch": 0.23089694582802425, + "grad_norm": 2.9860422611236572, + "learning_rate": 0.0001846094443800919, + "loss": 0.3436, + "step": 5980 + }, + { + "epoch": 0.23128306112205105, + "grad_norm": 13.12328815460205, + "learning_rate": 0.00018458370336049012, + "loss": 0.3952, + "step": 5990 + }, + { + "epoch": 0.23166917641607784, + "grad_norm": 3.7130823135375977, + "learning_rate": 0.00018455796234088834, + "loss": 0.3658, + "step": 6000 + }, + { + "epoch": 0.23205529171010464, + "grad_norm": 1.8329843282699585, + "learning_rate": 0.00018453222132128655, + "loss": 0.4172, + "step": 6010 + }, + { + "epoch": 0.23244140700413143, + "grad_norm": 1.3583799600601196, + "learning_rate": 0.00018450648030168476, + "loss": 0.4005, + "step": 6020 + }, + { + "epoch": 0.23282752229815823, + "grad_norm": 3.1711816787719727, + "learning_rate": 0.00018448073928208297, + "loss": 0.3674, + "step": 6030 + }, + { + "epoch": 0.23321363759218502, + "grad_norm": 1.576937198638916, + "learning_rate": 0.0001844549982624812, + "loss": 0.3444, + "step": 6040 + }, + { + "epoch": 0.23359975288621182, + "grad_norm": 3.922267436981201, + "learning_rate": 0.0001844292572428794, + "loss": 0.5939, + "step": 6050 + }, + { + "epoch": 0.23398586818023862, + "grad_norm": 2.9851067066192627, + "learning_rate": 0.00018440351622327761, + "loss": 0.2387, + "step": 6060 + }, + { + "epoch": 0.2343719834742654, + "grad_norm": 2.1216888427734375, + "learning_rate": 0.00018437777520367583, + "loss": 0.3836, + "step": 6070 + }, + { + "epoch": 0.2347580987682922, + "grad_norm": 2.9788095951080322, + "learning_rate": 0.00018435203418407404, + "loss": 0.474, + "step": 6080 + }, + { + "epoch": 0.235144214062319, + "grad_norm": 1.0204919576644897, + "learning_rate": 0.00018432629316447225, + "loss": 0.2837, + "step": 6090 + }, + { + "epoch": 0.2355303293563458, + "grad_norm": 0.9091696739196777, + "learning_rate": 0.00018430055214487047, + "loss": 0.6203, + "step": 6100 + }, + { + "epoch": 0.2359164446503726, + "grad_norm": 0.25899162888526917, + "learning_rate": 0.00018427481112526868, + "loss": 0.4759, + "step": 6110 + }, + { + "epoch": 0.2363025599443994, + "grad_norm": 1.8625538349151611, + "learning_rate": 0.0001842490701056669, + "loss": 0.2992, + "step": 6120 + }, + { + "epoch": 0.23668867523842618, + "grad_norm": 1.586521863937378, + "learning_rate": 0.0001842233290860651, + "loss": 0.6122, + "step": 6130 + }, + { + "epoch": 0.23707479053245298, + "grad_norm": 2.387650966644287, + "learning_rate": 0.00018419758806646332, + "loss": 0.3276, + "step": 6140 + }, + { + "epoch": 0.23746090582647977, + "grad_norm": 4.840515613555908, + "learning_rate": 0.00018417184704686153, + "loss": 0.6295, + "step": 6150 + }, + { + "epoch": 0.2378470211205066, + "grad_norm": 1.70024836063385, + "learning_rate": 0.00018414610602725975, + "loss": 0.2047, + "step": 6160 + }, + { + "epoch": 0.2382331364145334, + "grad_norm": 2.791619062423706, + "learning_rate": 0.00018412036500765796, + "loss": 0.4364, + "step": 6170 + }, + { + "epoch": 0.2386192517085602, + "grad_norm": 3.710066318511963, + "learning_rate": 0.00018409462398805617, + "loss": 0.4564, + "step": 6180 + }, + { + "epoch": 0.23900536700258698, + "grad_norm": 2.564347982406616, + "learning_rate": 0.0001840688829684544, + "loss": 0.3156, + "step": 6190 + }, + { + "epoch": 0.23939148229661378, + "grad_norm": 2.3921267986297607, + "learning_rate": 0.0001840431419488526, + "loss": 0.3483, + "step": 6200 + }, + { + "epoch": 0.23977759759064057, + "grad_norm": 1.4785810708999634, + "learning_rate": 0.00018401740092925081, + "loss": 0.4338, + "step": 6210 + }, + { + "epoch": 0.24016371288466737, + "grad_norm": 3.624790906906128, + "learning_rate": 0.00018399165990964903, + "loss": 0.7156, + "step": 6220 + }, + { + "epoch": 0.24054982817869416, + "grad_norm": 3.942161798477173, + "learning_rate": 0.00018396591889004724, + "loss": 0.3932, + "step": 6230 + }, + { + "epoch": 0.24093594347272096, + "grad_norm": 3.2236740589141846, + "learning_rate": 0.00018394017787044545, + "loss": 0.3933, + "step": 6240 + }, + { + "epoch": 0.24132205876674775, + "grad_norm": 2.5040500164031982, + "learning_rate": 0.00018391443685084367, + "loss": 0.5711, + "step": 6250 + }, + { + "epoch": 0.24170817406077455, + "grad_norm": 1.9934203624725342, + "learning_rate": 0.00018388869583124188, + "loss": 0.3074, + "step": 6260 + }, + { + "epoch": 0.24209428935480135, + "grad_norm": 3.702509641647339, + "learning_rate": 0.0001838629548116401, + "loss": 0.3454, + "step": 6270 + }, + { + "epoch": 0.24248040464882814, + "grad_norm": 2.076802968978882, + "learning_rate": 0.0001838372137920383, + "loss": 0.3044, + "step": 6280 + }, + { + "epoch": 0.24286651994285494, + "grad_norm": 5.798679351806641, + "learning_rate": 0.00018381147277243652, + "loss": 0.3396, + "step": 6290 + }, + { + "epoch": 0.24325263523688173, + "grad_norm": 4.698869705200195, + "learning_rate": 0.00018378573175283473, + "loss": 0.3735, + "step": 6300 + }, + { + "epoch": 0.24363875053090853, + "grad_norm": 3.029979705810547, + "learning_rate": 0.00018375999073323295, + "loss": 0.3891, + "step": 6310 + }, + { + "epoch": 0.24402486582493532, + "grad_norm": 2.5507185459136963, + "learning_rate": 0.00018373424971363116, + "loss": 0.4854, + "step": 6320 + }, + { + "epoch": 0.24441098111896212, + "grad_norm": 3.2052571773529053, + "learning_rate": 0.00018370850869402937, + "loss": 0.6789, + "step": 6330 + }, + { + "epoch": 0.2447970964129889, + "grad_norm": 1.9265435934066772, + "learning_rate": 0.00018368276767442761, + "loss": 0.4505, + "step": 6340 + }, + { + "epoch": 0.2451832117070157, + "grad_norm": 0.8391959071159363, + "learning_rate": 0.0001836570266548258, + "loss": 0.3432, + "step": 6350 + }, + { + "epoch": 0.2455693270010425, + "grad_norm": 3.4653851985931396, + "learning_rate": 0.00018363128563522401, + "loss": 0.3571, + "step": 6360 + }, + { + "epoch": 0.2459554422950693, + "grad_norm": 2.3033368587493896, + "learning_rate": 0.00018360554461562223, + "loss": 0.3625, + "step": 6370 + }, + { + "epoch": 0.2463415575890961, + "grad_norm": 1.659408450126648, + "learning_rate": 0.00018357980359602044, + "loss": 0.5311, + "step": 6380 + }, + { + "epoch": 0.2467276728831229, + "grad_norm": 1.1839714050292969, + "learning_rate": 0.00018355406257641865, + "loss": 0.3905, + "step": 6390 + }, + { + "epoch": 0.24711378817714968, + "grad_norm": 0.49230822920799255, + "learning_rate": 0.00018352832155681687, + "loss": 0.4021, + "step": 6400 + }, + { + "epoch": 0.24749990347117648, + "grad_norm": 4.451594829559326, + "learning_rate": 0.0001835025805372151, + "loss": 0.4504, + "step": 6410 + }, + { + "epoch": 0.2478860187652033, + "grad_norm": 1.0058324337005615, + "learning_rate": 0.0001834768395176133, + "loss": 0.2636, + "step": 6420 + }, + { + "epoch": 0.2482721340592301, + "grad_norm": 2.7853894233703613, + "learning_rate": 0.0001834510984980115, + "loss": 0.47, + "step": 6430 + }, + { + "epoch": 0.2486582493532569, + "grad_norm": 2.730095148086548, + "learning_rate": 0.00018342535747840972, + "loss": 0.3941, + "step": 6440 + }, + { + "epoch": 0.2490443646472837, + "grad_norm": 2.4993178844451904, + "learning_rate": 0.00018339961645880793, + "loss": 0.5777, + "step": 6450 + }, + { + "epoch": 0.24943047994131048, + "grad_norm": 2.361525297164917, + "learning_rate": 0.00018337387543920617, + "loss": 0.3798, + "step": 6460 + }, + { + "epoch": 0.24981659523533728, + "grad_norm": 2.5558526515960693, + "learning_rate": 0.00018334813441960436, + "loss": 0.3113, + "step": 6470 + }, + { + "epoch": 0.25020271052936405, + "grad_norm": 0.8033503890037537, + "learning_rate": 0.0001833223934000026, + "loss": 0.5254, + "step": 6480 + }, + { + "epoch": 0.25058882582339087, + "grad_norm": 2.721090078353882, + "learning_rate": 0.0001832966523804008, + "loss": 0.393, + "step": 6490 + }, + { + "epoch": 0.25097494111741764, + "grad_norm": 1.7147916555404663, + "learning_rate": 0.000183270911360799, + "loss": 0.3225, + "step": 6500 + }, + { + "epoch": 0.25136105641144446, + "grad_norm": 2.388347864151001, + "learning_rate": 0.00018324517034119721, + "loss": 0.3519, + "step": 6510 + }, + { + "epoch": 0.25174717170547123, + "grad_norm": 2.470891237258911, + "learning_rate": 0.00018321942932159543, + "loss": 0.4384, + "step": 6520 + }, + { + "epoch": 0.25213328699949805, + "grad_norm": 1.4743351936340332, + "learning_rate": 0.00018319368830199367, + "loss": 0.2464, + "step": 6530 + }, + { + "epoch": 0.2525194022935248, + "grad_norm": 1.5889122486114502, + "learning_rate": 0.00018316794728239185, + "loss": 0.3149, + "step": 6540 + }, + { + "epoch": 0.25290551758755164, + "grad_norm": 4.900819778442383, + "learning_rate": 0.0001831422062627901, + "loss": 0.3978, + "step": 6550 + }, + { + "epoch": 0.25329163288157847, + "grad_norm": 5.22566556930542, + "learning_rate": 0.00018311646524318828, + "loss": 0.4473, + "step": 6560 + }, + { + "epoch": 0.25367774817560523, + "grad_norm": 4.7480363845825195, + "learning_rate": 0.0001830907242235865, + "loss": 0.3976, + "step": 6570 + }, + { + "epoch": 0.25406386346963206, + "grad_norm": 1.4711374044418335, + "learning_rate": 0.0001830649832039847, + "loss": 0.5183, + "step": 6580 + }, + { + "epoch": 0.2544499787636588, + "grad_norm": 2.237309217453003, + "learning_rate": 0.00018303924218438292, + "loss": 0.2171, + "step": 6590 + }, + { + "epoch": 0.25483609405768565, + "grad_norm": 4.107303619384766, + "learning_rate": 0.00018301350116478116, + "loss": 0.3918, + "step": 6600 + }, + { + "epoch": 0.2552222093517124, + "grad_norm": 4.7285003662109375, + "learning_rate": 0.00018298776014517935, + "loss": 0.2042, + "step": 6610 + }, + { + "epoch": 0.25560832464573924, + "grad_norm": 2.1333792209625244, + "learning_rate": 0.0001829620191255776, + "loss": 0.3502, + "step": 6620 + }, + { + "epoch": 0.255994439939766, + "grad_norm": 3.062173843383789, + "learning_rate": 0.00018293627810597577, + "loss": 0.3949, + "step": 6630 + }, + { + "epoch": 0.25638055523379283, + "grad_norm": 1.538854956626892, + "learning_rate": 0.00018291053708637401, + "loss": 0.4613, + "step": 6640 + }, + { + "epoch": 0.2567666705278196, + "grad_norm": 2.546586751937866, + "learning_rate": 0.00018288479606677223, + "loss": 0.5868, + "step": 6650 + }, + { + "epoch": 0.2571527858218464, + "grad_norm": 2.7282049655914307, + "learning_rate": 0.00018285905504717041, + "loss": 0.4186, + "step": 6660 + }, + { + "epoch": 0.2575389011158732, + "grad_norm": 3.204634189605713, + "learning_rate": 0.00018283331402756865, + "loss": 0.4072, + "step": 6670 + }, + { + "epoch": 0.2579250164099, + "grad_norm": 2.421846866607666, + "learning_rate": 0.00018280757300796684, + "loss": 0.306, + "step": 6680 + }, + { + "epoch": 0.2583111317039268, + "grad_norm": 4.243416786193848, + "learning_rate": 0.00018278183198836508, + "loss": 0.2631, + "step": 6690 + }, + { + "epoch": 0.2586972469979536, + "grad_norm": 1.0495362281799316, + "learning_rate": 0.00018275609096876327, + "loss": 0.3488, + "step": 6700 + }, + { + "epoch": 0.25908336229198037, + "grad_norm": 1.915279746055603, + "learning_rate": 0.0001827303499491615, + "loss": 0.2589, + "step": 6710 + }, + { + "epoch": 0.2594694775860072, + "grad_norm": 3.724299192428589, + "learning_rate": 0.00018270460892955972, + "loss": 0.5118, + "step": 6720 + }, + { + "epoch": 0.25985559288003396, + "grad_norm": 2.832204580307007, + "learning_rate": 0.0001826788679099579, + "loss": 0.2508, + "step": 6730 + }, + { + "epoch": 0.2602417081740608, + "grad_norm": 1.1942508220672607, + "learning_rate": 0.00018265312689035615, + "loss": 0.4328, + "step": 6740 + }, + { + "epoch": 0.26062782346808755, + "grad_norm": 1.0741711854934692, + "learning_rate": 0.00018262738587075433, + "loss": 0.3514, + "step": 6750 + }, + { + "epoch": 0.2610139387621144, + "grad_norm": 2.9918277263641357, + "learning_rate": 0.00018260164485115257, + "loss": 0.3528, + "step": 6760 + }, + { + "epoch": 0.26140005405614114, + "grad_norm": 1.3773655891418457, + "learning_rate": 0.0001825759038315508, + "loss": 0.365, + "step": 6770 + }, + { + "epoch": 0.26178616935016796, + "grad_norm": 3.5288615226745605, + "learning_rate": 0.000182550162811949, + "loss": 0.3645, + "step": 6780 + }, + { + "epoch": 0.26217228464419473, + "grad_norm": 1.2178785800933838, + "learning_rate": 0.00018252442179234721, + "loss": 0.3742, + "step": 6790 + }, + { + "epoch": 0.26255839993822155, + "grad_norm": 2.7981081008911133, + "learning_rate": 0.0001824986807727454, + "loss": 0.6174, + "step": 6800 + }, + { + "epoch": 0.2629445152322484, + "grad_norm": 1.6766215562820435, + "learning_rate": 0.00018247293975314364, + "loss": 0.3028, + "step": 6810 + }, + { + "epoch": 0.26333063052627514, + "grad_norm": 3.7797629833221436, + "learning_rate": 0.00018244719873354183, + "loss": 0.2633, + "step": 6820 + }, + { + "epoch": 0.26371674582030197, + "grad_norm": 7.794743537902832, + "learning_rate": 0.00018242145771394007, + "loss": 0.3586, + "step": 6830 + }, + { + "epoch": 0.26410286111432874, + "grad_norm": 0.5704814195632935, + "learning_rate": 0.00018239571669433828, + "loss": 0.3506, + "step": 6840 + }, + { + "epoch": 0.26448897640835556, + "grad_norm": 5.771059513092041, + "learning_rate": 0.0001823699756747365, + "loss": 0.3881, + "step": 6850 + }, + { + "epoch": 0.2648750917023823, + "grad_norm": 2.723592519760132, + "learning_rate": 0.0001823442346551347, + "loss": 0.3955, + "step": 6860 + }, + { + "epoch": 0.26526120699640915, + "grad_norm": 1.5448215007781982, + "learning_rate": 0.0001823184936355329, + "loss": 0.495, + "step": 6870 + }, + { + "epoch": 0.2656473222904359, + "grad_norm": 2.2980363368988037, + "learning_rate": 0.00018229275261593113, + "loss": 0.2695, + "step": 6880 + }, + { + "epoch": 0.26603343758446274, + "grad_norm": 1.959811806678772, + "learning_rate": 0.00018226701159632932, + "loss": 0.383, + "step": 6890 + }, + { + "epoch": 0.2664195528784895, + "grad_norm": 2.1491482257843018, + "learning_rate": 0.00018224127057672756, + "loss": 0.5655, + "step": 6900 + }, + { + "epoch": 0.26680566817251633, + "grad_norm": 6.472841262817383, + "learning_rate": 0.00018221552955712577, + "loss": 0.4757, + "step": 6910 + }, + { + "epoch": 0.2671917834665431, + "grad_norm": 7.878561496734619, + "learning_rate": 0.000182189788537524, + "loss": 0.3944, + "step": 6920 + }, + { + "epoch": 0.2675778987605699, + "grad_norm": 0.052701435983181, + "learning_rate": 0.0001821640475179222, + "loss": 0.382, + "step": 6930 + }, + { + "epoch": 0.2679640140545967, + "grad_norm": 2.294677972793579, + "learning_rate": 0.00018213830649832039, + "loss": 0.2932, + "step": 6940 + }, + { + "epoch": 0.2683501293486235, + "grad_norm": 1.6058757305145264, + "learning_rate": 0.00018211256547871863, + "loss": 0.4438, + "step": 6950 + }, + { + "epoch": 0.2687362446426503, + "grad_norm": 4.003495693206787, + "learning_rate": 0.00018208682445911684, + "loss": 0.5945, + "step": 6960 + }, + { + "epoch": 0.2691223599366771, + "grad_norm": 1.423017144203186, + "learning_rate": 0.00018206108343951505, + "loss": 0.4356, + "step": 6970 + }, + { + "epoch": 0.26950847523070387, + "grad_norm": 2.206341028213501, + "learning_rate": 0.00018203534241991327, + "loss": 0.344, + "step": 6980 + }, + { + "epoch": 0.2698945905247307, + "grad_norm": 0.6644784212112427, + "learning_rate": 0.00018200960140031148, + "loss": 0.4988, + "step": 6990 + }, + { + "epoch": 0.27028070581875746, + "grad_norm": 2.4569833278656006, + "learning_rate": 0.0001819838603807097, + "loss": 0.3689, + "step": 7000 + }, + { + "epoch": 0.2706668211127843, + "grad_norm": 1.554567575454712, + "learning_rate": 0.00018195811936110788, + "loss": 0.4684, + "step": 7010 + }, + { + "epoch": 0.27105293640681105, + "grad_norm": 3.2556328773498535, + "learning_rate": 0.00018193237834150612, + "loss": 0.611, + "step": 7020 + }, + { + "epoch": 0.2714390517008379, + "grad_norm": 2.9123427867889404, + "learning_rate": 0.00018190663732190433, + "loss": 0.4278, + "step": 7030 + }, + { + "epoch": 0.27182516699486464, + "grad_norm": 2.159273862838745, + "learning_rate": 0.00018188089630230255, + "loss": 0.2384, + "step": 7040 + }, + { + "epoch": 0.27221128228889147, + "grad_norm": 3.4977822303771973, + "learning_rate": 0.00018185515528270076, + "loss": 0.5459, + "step": 7050 + }, + { + "epoch": 0.27259739758291823, + "grad_norm": 1.1822031736373901, + "learning_rate": 0.00018182941426309897, + "loss": 0.4364, + "step": 7060 + }, + { + "epoch": 0.27298351287694506, + "grad_norm": 2.4467339515686035, + "learning_rate": 0.00018180367324349719, + "loss": 0.5198, + "step": 7070 + }, + { + "epoch": 0.2733696281709719, + "grad_norm": 1.0406467914581299, + "learning_rate": 0.0001817779322238954, + "loss": 0.2797, + "step": 7080 + }, + { + "epoch": 0.27375574346499865, + "grad_norm": 1.925830602645874, + "learning_rate": 0.0001817521912042936, + "loss": 0.4898, + "step": 7090 + }, + { + "epoch": 0.27414185875902547, + "grad_norm": 3.0385682582855225, + "learning_rate": 0.00018172645018469183, + "loss": 0.3867, + "step": 7100 + }, + { + "epoch": 0.27452797405305224, + "grad_norm": 1.5285695791244507, + "learning_rate": 0.00018170070916509004, + "loss": 0.4233, + "step": 7110 + }, + { + "epoch": 0.27491408934707906, + "grad_norm": 1.266693115234375, + "learning_rate": 0.00018167496814548825, + "loss": 0.4724, + "step": 7120 + }, + { + "epoch": 0.27530020464110583, + "grad_norm": 3.371323585510254, + "learning_rate": 0.00018164922712588647, + "loss": 0.533, + "step": 7130 + }, + { + "epoch": 0.27568631993513265, + "grad_norm": 2.662691116333008, + "learning_rate": 0.00018162348610628468, + "loss": 0.3134, + "step": 7140 + }, + { + "epoch": 0.2760724352291594, + "grad_norm": 1.8977057933807373, + "learning_rate": 0.0001815977450866829, + "loss": 0.3038, + "step": 7150 + }, + { + "epoch": 0.27645855052318624, + "grad_norm": 3.1027894020080566, + "learning_rate": 0.0001815720040670811, + "loss": 0.5074, + "step": 7160 + }, + { + "epoch": 0.276844665817213, + "grad_norm": 1.2112785577774048, + "learning_rate": 0.00018154626304747932, + "loss": 0.324, + "step": 7170 + }, + { + "epoch": 0.27723078111123983, + "grad_norm": 1.6500996351242065, + "learning_rate": 0.00018152052202787753, + "loss": 0.2856, + "step": 7180 + }, + { + "epoch": 0.2776168964052666, + "grad_norm": 3.215747833251953, + "learning_rate": 0.00018149478100827575, + "loss": 0.4522, + "step": 7190 + }, + { + "epoch": 0.2780030116992934, + "grad_norm": 4.8541059494018555, + "learning_rate": 0.00018146903998867396, + "loss": 0.4106, + "step": 7200 + }, + { + "epoch": 0.2783891269933202, + "grad_norm": 2.3697152137756348, + "learning_rate": 0.00018144329896907217, + "loss": 0.2673, + "step": 7210 + }, + { + "epoch": 0.278775242287347, + "grad_norm": 2.9693639278411865, + "learning_rate": 0.00018141755794947039, + "loss": 0.3949, + "step": 7220 + }, + { + "epoch": 0.2791613575813738, + "grad_norm": 2.691817283630371, + "learning_rate": 0.0001813918169298686, + "loss": 0.3427, + "step": 7230 + }, + { + "epoch": 0.2795474728754006, + "grad_norm": 5.197331428527832, + "learning_rate": 0.0001813660759102668, + "loss": 0.4331, + "step": 7240 + }, + { + "epoch": 0.27993358816942737, + "grad_norm": 1.5799933671951294, + "learning_rate": 0.00018134033489066503, + "loss": 0.3543, + "step": 7250 + }, + { + "epoch": 0.2803197034634542, + "grad_norm": 1.3614271879196167, + "learning_rate": 0.00018131459387106324, + "loss": 0.5289, + "step": 7260 + }, + { + "epoch": 0.28070581875748096, + "grad_norm": 2.2942802906036377, + "learning_rate": 0.00018128885285146145, + "loss": 0.4318, + "step": 7270 + }, + { + "epoch": 0.2810919340515078, + "grad_norm": 1.1805604696273804, + "learning_rate": 0.00018126311183185967, + "loss": 0.4754, + "step": 7280 + }, + { + "epoch": 0.28147804934553455, + "grad_norm": 0.5108867883682251, + "learning_rate": 0.00018123737081225788, + "loss": 0.4517, + "step": 7290 + }, + { + "epoch": 0.2818641646395614, + "grad_norm": 1.1736596822738647, + "learning_rate": 0.0001812116297926561, + "loss": 0.4538, + "step": 7300 + }, + { + "epoch": 0.28225027993358814, + "grad_norm": 5.497414588928223, + "learning_rate": 0.0001811858887730543, + "loss": 0.5116, + "step": 7310 + }, + { + "epoch": 0.28263639522761497, + "grad_norm": 1.1347368955612183, + "learning_rate": 0.00018116014775345252, + "loss": 0.3848, + "step": 7320 + }, + { + "epoch": 0.28302251052164173, + "grad_norm": 2.740715742111206, + "learning_rate": 0.00018113440673385073, + "loss": 0.3456, + "step": 7330 + }, + { + "epoch": 0.28340862581566856, + "grad_norm": 1.3853389024734497, + "learning_rate": 0.00018110866571424897, + "loss": 0.3398, + "step": 7340 + }, + { + "epoch": 0.2837947411096954, + "grad_norm": 7.493706703186035, + "learning_rate": 0.00018108292469464716, + "loss": 0.2726, + "step": 7350 + }, + { + "epoch": 0.28418085640372215, + "grad_norm": 1.81704843044281, + "learning_rate": 0.00018105718367504537, + "loss": 0.3818, + "step": 7360 + }, + { + "epoch": 0.28456697169774897, + "grad_norm": 2.4877755641937256, + "learning_rate": 0.00018103144265544359, + "loss": 0.3499, + "step": 7370 + }, + { + "epoch": 0.28495308699177574, + "grad_norm": 1.3704471588134766, + "learning_rate": 0.0001810057016358418, + "loss": 0.2346, + "step": 7380 + }, + { + "epoch": 0.28533920228580256, + "grad_norm": 2.664745569229126, + "learning_rate": 0.00018097996061624, + "loss": 0.4041, + "step": 7390 + }, + { + "epoch": 0.28572531757982933, + "grad_norm": 3.6539089679718018, + "learning_rate": 0.00018095421959663823, + "loss": 0.2885, + "step": 7400 + }, + { + "epoch": 0.28611143287385615, + "grad_norm": 0.8653857707977295, + "learning_rate": 0.00018092847857703647, + "loss": 0.3849, + "step": 7410 + }, + { + "epoch": 0.2864975481678829, + "grad_norm": 2.6319446563720703, + "learning_rate": 0.00018090273755743465, + "loss": 0.2728, + "step": 7420 + }, + { + "epoch": 0.28688366346190974, + "grad_norm": 2.3457818031311035, + "learning_rate": 0.00018087699653783287, + "loss": 0.446, + "step": 7430 + }, + { + "epoch": 0.2872697787559365, + "grad_norm": 0.8546158671379089, + "learning_rate": 0.00018085125551823108, + "loss": 0.2898, + "step": 7440 + }, + { + "epoch": 0.28765589404996333, + "grad_norm": 0.45937278866767883, + "learning_rate": 0.0001808255144986293, + "loss": 0.583, + "step": 7450 + }, + { + "epoch": 0.2880420093439901, + "grad_norm": 1.7129520177841187, + "learning_rate": 0.00018079977347902753, + "loss": 0.4908, + "step": 7460 + }, + { + "epoch": 0.2884281246380169, + "grad_norm": 4.106715679168701, + "learning_rate": 0.00018077403245942572, + "loss": 0.3373, + "step": 7470 + }, + { + "epoch": 0.2888142399320437, + "grad_norm": 3.8112800121307373, + "learning_rate": 0.00018074829143982396, + "loss": 0.392, + "step": 7480 + }, + { + "epoch": 0.2892003552260705, + "grad_norm": 0.5382593274116516, + "learning_rate": 0.00018072255042022215, + "loss": 0.2929, + "step": 7490 + }, + { + "epoch": 0.2895864705200973, + "grad_norm": 2.50888991355896, + "learning_rate": 0.00018069680940062036, + "loss": 0.3361, + "step": 7500 + }, + { + "epoch": 0.2899725858141241, + "grad_norm": 3.3544275760650635, + "learning_rate": 0.00018067106838101857, + "loss": 0.388, + "step": 7510 + }, + { + "epoch": 0.2903587011081509, + "grad_norm": 1.192386507987976, + "learning_rate": 0.00018064532736141679, + "loss": 0.4427, + "step": 7520 + }, + { + "epoch": 0.2907448164021777, + "grad_norm": 1.5527079105377197, + "learning_rate": 0.00018061958634181503, + "loss": 0.4023, + "step": 7530 + }, + { + "epoch": 0.29113093169620446, + "grad_norm": 0.67446368932724, + "learning_rate": 0.0001805938453222132, + "loss": 0.4949, + "step": 7540 + }, + { + "epoch": 0.2915170469902313, + "grad_norm": 1.6349838972091675, + "learning_rate": 0.00018056810430261145, + "loss": 0.3811, + "step": 7550 + }, + { + "epoch": 0.29190316228425806, + "grad_norm": 1.4848904609680176, + "learning_rate": 0.00018054236328300964, + "loss": 0.3851, + "step": 7560 + }, + { + "epoch": 0.2922892775782849, + "grad_norm": 0.9933151006698608, + "learning_rate": 0.00018051662226340785, + "loss": 0.4699, + "step": 7570 + }, + { + "epoch": 0.29267539287231165, + "grad_norm": 1.1026233434677124, + "learning_rate": 0.00018049088124380607, + "loss": 0.3287, + "step": 7580 + }, + { + "epoch": 0.29306150816633847, + "grad_norm": 1.232954740524292, + "learning_rate": 0.00018046514022420428, + "loss": 0.3722, + "step": 7590 + }, + { + "epoch": 0.2934476234603653, + "grad_norm": 3.8303146362304688, + "learning_rate": 0.00018043939920460252, + "loss": 0.2985, + "step": 7600 + }, + { + "epoch": 0.29383373875439206, + "grad_norm": 1.9358845949172974, + "learning_rate": 0.0001804136581850007, + "loss": 0.4361, + "step": 7610 + }, + { + "epoch": 0.2942198540484189, + "grad_norm": 1.8905962705612183, + "learning_rate": 0.00018038791716539895, + "loss": 0.2835, + "step": 7620 + }, + { + "epoch": 0.29460596934244565, + "grad_norm": 1.9965651035308838, + "learning_rate": 0.00018036217614579713, + "loss": 0.5387, + "step": 7630 + }, + { + "epoch": 0.2949920846364725, + "grad_norm": 4.204270839691162, + "learning_rate": 0.00018033643512619535, + "loss": 0.3498, + "step": 7640 + }, + { + "epoch": 0.29537819993049924, + "grad_norm": 1.4732340574264526, + "learning_rate": 0.00018031069410659359, + "loss": 0.315, + "step": 7650 + }, + { + "epoch": 0.29576431522452606, + "grad_norm": 1.0233594179153442, + "learning_rate": 0.00018028495308699177, + "loss": 0.1536, + "step": 7660 + }, + { + "epoch": 0.29615043051855283, + "grad_norm": 3.1531457901000977, + "learning_rate": 0.00018025921206739, + "loss": 0.3793, + "step": 7670 + }, + { + "epoch": 0.29653654581257965, + "grad_norm": 0.8080945014953613, + "learning_rate": 0.0001802334710477882, + "loss": 0.5589, + "step": 7680 + }, + { + "epoch": 0.2969226611066064, + "grad_norm": 3.1202728748321533, + "learning_rate": 0.00018020773002818644, + "loss": 0.4652, + "step": 7690 + }, + { + "epoch": 0.29730877640063325, + "grad_norm": 2.5934784412384033, + "learning_rate": 0.00018018198900858463, + "loss": 0.4921, + "step": 7700 + }, + { + "epoch": 0.29769489169466, + "grad_norm": 2.858642101287842, + "learning_rate": 0.00018015624798898284, + "loss": 0.2732, + "step": 7710 + }, + { + "epoch": 0.29808100698868684, + "grad_norm": 3.621229887008667, + "learning_rate": 0.00018013050696938108, + "loss": 0.5639, + "step": 7720 + }, + { + "epoch": 0.2984671222827136, + "grad_norm": 3.7943220138549805, + "learning_rate": 0.00018010476594977926, + "loss": 0.3177, + "step": 7730 + }, + { + "epoch": 0.2988532375767404, + "grad_norm": 1.6371623277664185, + "learning_rate": 0.0001800790249301775, + "loss": 0.4211, + "step": 7740 + }, + { + "epoch": 0.2992393528707672, + "grad_norm": 1.9557713270187378, + "learning_rate": 0.0001800532839105757, + "loss": 0.4351, + "step": 7750 + }, + { + "epoch": 0.299625468164794, + "grad_norm": 2.684964895248413, + "learning_rate": 0.00018002754289097393, + "loss": 0.39, + "step": 7760 + }, + { + "epoch": 0.3000115834588208, + "grad_norm": 1.7401316165924072, + "learning_rate": 0.00018000180187137215, + "loss": 0.2844, + "step": 7770 + }, + { + "epoch": 0.3003976987528476, + "grad_norm": 0.6305844187736511, + "learning_rate": 0.00017997606085177033, + "loss": 0.2472, + "step": 7780 + }, + { + "epoch": 0.3007838140468744, + "grad_norm": 2.2880289554595947, + "learning_rate": 0.00017995031983216857, + "loss": 0.3952, + "step": 7790 + }, + { + "epoch": 0.3011699293409012, + "grad_norm": 3.423980951309204, + "learning_rate": 0.00017992457881256676, + "loss": 0.4459, + "step": 7800 + }, + { + "epoch": 0.30155604463492797, + "grad_norm": 0.6920475363731384, + "learning_rate": 0.000179898837792965, + "loss": 0.2909, + "step": 7810 + }, + { + "epoch": 0.3019421599289548, + "grad_norm": 0.8905349373817444, + "learning_rate": 0.00017987309677336318, + "loss": 0.346, + "step": 7820 + }, + { + "epoch": 0.30232827522298156, + "grad_norm": 1.8836702108383179, + "learning_rate": 0.00017984735575376143, + "loss": 0.4038, + "step": 7830 + }, + { + "epoch": 0.3027143905170084, + "grad_norm": 2.6712753772735596, + "learning_rate": 0.00017982161473415964, + "loss": 0.3452, + "step": 7840 + }, + { + "epoch": 0.30310050581103515, + "grad_norm": 2.344122886657715, + "learning_rate": 0.00017979587371455785, + "loss": 0.5091, + "step": 7850 + }, + { + "epoch": 0.30348662110506197, + "grad_norm": 3.734415054321289, + "learning_rate": 0.00017977013269495607, + "loss": 0.3893, + "step": 7860 + }, + { + "epoch": 0.3038727363990888, + "grad_norm": 1.70572829246521, + "learning_rate": 0.00017974439167535425, + "loss": 0.4829, + "step": 7870 + }, + { + "epoch": 0.30425885169311556, + "grad_norm": 1.779189109802246, + "learning_rate": 0.0001797186506557525, + "loss": 0.5361, + "step": 7880 + }, + { + "epoch": 0.3046449669871424, + "grad_norm": 2.888803482055664, + "learning_rate": 0.00017969290963615068, + "loss": 0.4305, + "step": 7890 + }, + { + "epoch": 0.30503108228116915, + "grad_norm": 1.2247655391693115, + "learning_rate": 0.00017966716861654892, + "loss": 0.3817, + "step": 7900 + }, + { + "epoch": 0.305417197575196, + "grad_norm": 2.995152473449707, + "learning_rate": 0.00017964142759694713, + "loss": 0.4669, + "step": 7910 + }, + { + "epoch": 0.30580331286922274, + "grad_norm": 8.049060821533203, + "learning_rate": 0.00017961568657734535, + "loss": 0.6706, + "step": 7920 + }, + { + "epoch": 0.30618942816324957, + "grad_norm": 2.1181435585021973, + "learning_rate": 0.00017958994555774356, + "loss": 0.4353, + "step": 7930 + }, + { + "epoch": 0.30657554345727633, + "grad_norm": 8.394509315490723, + "learning_rate": 0.00017956420453814174, + "loss": 0.3497, + "step": 7940 + }, + { + "epoch": 0.30696165875130316, + "grad_norm": 2.5140750408172607, + "learning_rate": 0.00017953846351853998, + "loss": 0.5774, + "step": 7950 + }, + { + "epoch": 0.3073477740453299, + "grad_norm": 2.720942974090576, + "learning_rate": 0.0001795127224989382, + "loss": 0.4457, + "step": 7960 + }, + { + "epoch": 0.30773388933935675, + "grad_norm": 1.8155667781829834, + "learning_rate": 0.0001794869814793364, + "loss": 0.4155, + "step": 7970 + }, + { + "epoch": 0.3081200046333835, + "grad_norm": 1.9989752769470215, + "learning_rate": 0.00017946124045973462, + "loss": 0.3233, + "step": 7980 + }, + { + "epoch": 0.30850611992741034, + "grad_norm": 0.7483557462692261, + "learning_rate": 0.00017943549944013284, + "loss": 0.2932, + "step": 7990 + }, + { + "epoch": 0.3088922352214371, + "grad_norm": 0.5750642418861389, + "learning_rate": 0.00017940975842053105, + "loss": 0.401, + "step": 8000 + }, + { + "epoch": 0.30927835051546393, + "grad_norm": 1.2084500789642334, + "learning_rate": 0.00017938401740092924, + "loss": 0.3705, + "step": 8010 + }, + { + "epoch": 0.3096644658094907, + "grad_norm": 1.833434820175171, + "learning_rate": 0.00017935827638132748, + "loss": 0.3507, + "step": 8020 + }, + { + "epoch": 0.3100505811035175, + "grad_norm": 3.147508382797241, + "learning_rate": 0.0001793325353617257, + "loss": 0.3255, + "step": 8030 + }, + { + "epoch": 0.3104366963975443, + "grad_norm": 2.150932788848877, + "learning_rate": 0.0001793067943421239, + "loss": 0.3401, + "step": 8040 + }, + { + "epoch": 0.3108228116915711, + "grad_norm": 3.3340635299682617, + "learning_rate": 0.00017928105332252212, + "loss": 0.3606, + "step": 8050 + }, + { + "epoch": 0.3112089269855979, + "grad_norm": 5.173205375671387, + "learning_rate": 0.00017925531230292033, + "loss": 0.1695, + "step": 8060 + }, + { + "epoch": 0.3115950422796247, + "grad_norm": 1.0863877534866333, + "learning_rate": 0.00017922957128331854, + "loss": 0.3038, + "step": 8070 + }, + { + "epoch": 0.31198115757365147, + "grad_norm": 1.5977118015289307, + "learning_rate": 0.00017920383026371676, + "loss": 0.2291, + "step": 8080 + }, + { + "epoch": 0.3123672728676783, + "grad_norm": 4.040243625640869, + "learning_rate": 0.00017917808924411497, + "loss": 0.8538, + "step": 8090 + }, + { + "epoch": 0.31275338816170506, + "grad_norm": 1.5926854610443115, + "learning_rate": 0.00017915234822451318, + "loss": 0.4733, + "step": 8100 + }, + { + "epoch": 0.3131395034557319, + "grad_norm": 1.0959421396255493, + "learning_rate": 0.0001791266072049114, + "loss": 0.6076, + "step": 8110 + }, + { + "epoch": 0.3135256187497587, + "grad_norm": 2.786085367202759, + "learning_rate": 0.0001791008661853096, + "loss": 0.3229, + "step": 8120 + }, + { + "epoch": 0.3139117340437855, + "grad_norm": 2.2573914527893066, + "learning_rate": 0.00017907512516570782, + "loss": 0.3676, + "step": 8130 + }, + { + "epoch": 0.3142978493378123, + "grad_norm": 2.271852493286133, + "learning_rate": 0.00017904938414610604, + "loss": 0.6275, + "step": 8140 + }, + { + "epoch": 0.31468396463183906, + "grad_norm": 1.9762821197509766, + "learning_rate": 0.00017902364312650425, + "loss": 0.2232, + "step": 8150 + }, + { + "epoch": 0.3150700799258659, + "grad_norm": 2.9960873126983643, + "learning_rate": 0.00017899790210690246, + "loss": 0.4739, + "step": 8160 + }, + { + "epoch": 0.31545619521989265, + "grad_norm": 1.142216682434082, + "learning_rate": 0.00017897216108730068, + "loss": 0.5983, + "step": 8170 + }, + { + "epoch": 0.3158423105139195, + "grad_norm": 1.7127768993377686, + "learning_rate": 0.0001789464200676989, + "loss": 0.4131, + "step": 8180 + }, + { + "epoch": 0.31622842580794625, + "grad_norm": 1.579793095588684, + "learning_rate": 0.0001789206790480971, + "loss": 0.3119, + "step": 8190 + }, + { + "epoch": 0.31661454110197307, + "grad_norm": 0.9647886157035828, + "learning_rate": 0.00017889493802849532, + "loss": 0.398, + "step": 8200 + }, + { + "epoch": 0.31700065639599984, + "grad_norm": 3.435312032699585, + "learning_rate": 0.00017886919700889353, + "loss": 0.405, + "step": 8210 + }, + { + "epoch": 0.31738677169002666, + "grad_norm": 2.1500205993652344, + "learning_rate": 0.00017884345598929174, + "loss": 0.2519, + "step": 8220 + }, + { + "epoch": 0.3177728869840534, + "grad_norm": 1.3107216358184814, + "learning_rate": 0.00017881771496968996, + "loss": 0.2846, + "step": 8230 + }, + { + "epoch": 0.31815900227808025, + "grad_norm": 0.1899029165506363, + "learning_rate": 0.00017879197395008817, + "loss": 0.4597, + "step": 8240 + }, + { + "epoch": 0.318545117572107, + "grad_norm": 2.329299211502075, + "learning_rate": 0.00017876623293048638, + "loss": 0.6523, + "step": 8250 + }, + { + "epoch": 0.31893123286613384, + "grad_norm": 0.41523978114128113, + "learning_rate": 0.0001787404919108846, + "loss": 0.3266, + "step": 8260 + }, + { + "epoch": 0.3193173481601606, + "grad_norm": 0.7914639711380005, + "learning_rate": 0.0001787147508912828, + "loss": 0.4029, + "step": 8270 + }, + { + "epoch": 0.31970346345418743, + "grad_norm": 0.6159287691116333, + "learning_rate": 0.00017868900987168102, + "loss": 0.4426, + "step": 8280 + }, + { + "epoch": 0.3200895787482142, + "grad_norm": 1.3690640926361084, + "learning_rate": 0.00017866326885207924, + "loss": 0.2974, + "step": 8290 + }, + { + "epoch": 0.320475694042241, + "grad_norm": 0.8592869639396667, + "learning_rate": 0.00017863752783247745, + "loss": 0.232, + "step": 8300 + }, + { + "epoch": 0.3208618093362678, + "grad_norm": 0.43169018626213074, + "learning_rate": 0.00017861178681287566, + "loss": 0.4033, + "step": 8310 + }, + { + "epoch": 0.3212479246302946, + "grad_norm": 0.8405828475952148, + "learning_rate": 0.00017858604579327388, + "loss": 0.3339, + "step": 8320 + }, + { + "epoch": 0.3216340399243214, + "grad_norm": 2.3412604331970215, + "learning_rate": 0.0001785603047736721, + "loss": 0.2781, + "step": 8330 + }, + { + "epoch": 0.3220201552183482, + "grad_norm": 2.412045478820801, + "learning_rate": 0.0001785345637540703, + "loss": 0.4346, + "step": 8340 + }, + { + "epoch": 0.32240627051237497, + "grad_norm": 3.626305341720581, + "learning_rate": 0.00017850882273446852, + "loss": 0.327, + "step": 8350 + }, + { + "epoch": 0.3227923858064018, + "grad_norm": 0.5645825266838074, + "learning_rate": 0.00017848308171486673, + "loss": 0.234, + "step": 8360 + }, + { + "epoch": 0.32317850110042856, + "grad_norm": 4.27307653427124, + "learning_rate": 0.00017845734069526494, + "loss": 0.5493, + "step": 8370 + }, + { + "epoch": 0.3235646163944554, + "grad_norm": 0.4511154890060425, + "learning_rate": 0.00017843159967566316, + "loss": 0.3501, + "step": 8380 + }, + { + "epoch": 0.3239507316884822, + "grad_norm": 0.314996600151062, + "learning_rate": 0.00017840585865606137, + "loss": 0.3544, + "step": 8390 + }, + { + "epoch": 0.324336846982509, + "grad_norm": 1.6546530723571777, + "learning_rate": 0.00017838011763645958, + "loss": 0.2455, + "step": 8400 + }, + { + "epoch": 0.3247229622765358, + "grad_norm": 3.2812252044677734, + "learning_rate": 0.0001783543766168578, + "loss": 0.3333, + "step": 8410 + }, + { + "epoch": 0.32510907757056257, + "grad_norm": 3.5717616081237793, + "learning_rate": 0.000178328635597256, + "loss": 0.4679, + "step": 8420 + }, + { + "epoch": 0.3254951928645894, + "grad_norm": 1.12017023563385, + "learning_rate": 0.00017830289457765422, + "loss": 0.3481, + "step": 8430 + }, + { + "epoch": 0.32588130815861616, + "grad_norm": 1.869462490081787, + "learning_rate": 0.00017827715355805244, + "loss": 0.4566, + "step": 8440 + }, + { + "epoch": 0.326267423452643, + "grad_norm": 1.4613149166107178, + "learning_rate": 0.00017825141253845065, + "loss": 0.5456, + "step": 8450 + }, + { + "epoch": 0.32665353874666975, + "grad_norm": 0.6842670440673828, + "learning_rate": 0.0001782256715188489, + "loss": 0.2776, + "step": 8460 + }, + { + "epoch": 0.32703965404069657, + "grad_norm": 2.9485504627227783, + "learning_rate": 0.00017819993049924708, + "loss": 0.3204, + "step": 8470 + }, + { + "epoch": 0.32742576933472334, + "grad_norm": 3.2084853649139404, + "learning_rate": 0.0001781741894796453, + "loss": 0.3391, + "step": 8480 + }, + { + "epoch": 0.32781188462875016, + "grad_norm": 1.9732774496078491, + "learning_rate": 0.0001781484484600435, + "loss": 0.3283, + "step": 8490 + }, + { + "epoch": 0.32819799992277693, + "grad_norm": 0.6378610134124756, + "learning_rate": 0.00017812270744044172, + "loss": 0.4519, + "step": 8500 + }, + { + "epoch": 0.32858411521680375, + "grad_norm": 4.108947277069092, + "learning_rate": 0.00017809696642083993, + "loss": 0.4933, + "step": 8510 + }, + { + "epoch": 0.3289702305108305, + "grad_norm": 2.7623212337493896, + "learning_rate": 0.00017807122540123814, + "loss": 0.4197, + "step": 8520 + }, + { + "epoch": 0.32935634580485734, + "grad_norm": 1.8904645442962646, + "learning_rate": 0.00017804548438163638, + "loss": 0.325, + "step": 8530 + }, + { + "epoch": 0.3297424610988841, + "grad_norm": 0.5131659507751465, + "learning_rate": 0.00017801974336203457, + "loss": 0.2403, + "step": 8540 + }, + { + "epoch": 0.33012857639291093, + "grad_norm": 2.965916633605957, + "learning_rate": 0.0001779940023424328, + "loss": 0.4597, + "step": 8550 + }, + { + "epoch": 0.3305146916869377, + "grad_norm": 1.5409698486328125, + "learning_rate": 0.000177968261322831, + "loss": 0.4594, + "step": 8560 + }, + { + "epoch": 0.3309008069809645, + "grad_norm": 1.1746805906295776, + "learning_rate": 0.0001779425203032292, + "loss": 0.4581, + "step": 8570 + }, + { + "epoch": 0.3312869222749913, + "grad_norm": 4.493356227874756, + "learning_rate": 0.00017791677928362745, + "loss": 0.4699, + "step": 8580 + }, + { + "epoch": 0.3316730375690181, + "grad_norm": 3.506526470184326, + "learning_rate": 0.00017789103826402564, + "loss": 0.3974, + "step": 8590 + }, + { + "epoch": 0.3320591528630449, + "grad_norm": 2.3893234729766846, + "learning_rate": 0.00017786529724442388, + "loss": 0.2823, + "step": 8600 + }, + { + "epoch": 0.3324452681570717, + "grad_norm": 1.6228163242340088, + "learning_rate": 0.00017783955622482206, + "loss": 0.4199, + "step": 8610 + }, + { + "epoch": 0.3328313834510985, + "grad_norm": 3.2869131565093994, + "learning_rate": 0.0001778138152052203, + "loss": 0.3173, + "step": 8620 + }, + { + "epoch": 0.3332174987451253, + "grad_norm": 5.547116279602051, + "learning_rate": 0.0001777880741856185, + "loss": 0.4584, + "step": 8630 + }, + { + "epoch": 0.33360361403915206, + "grad_norm": 1.3338594436645508, + "learning_rate": 0.0001777623331660167, + "loss": 0.4235, + "step": 8640 + }, + { + "epoch": 0.3339897293331789, + "grad_norm": 1.9165093898773193, + "learning_rate": 0.00017773659214641494, + "loss": 0.2989, + "step": 8650 + }, + { + "epoch": 0.3343758446272057, + "grad_norm": 1.968935251235962, + "learning_rate": 0.00017771085112681313, + "loss": 0.4194, + "step": 8660 + }, + { + "epoch": 0.3347619599212325, + "grad_norm": 9.66997241973877, + "learning_rate": 0.00017768511010721137, + "loss": 0.5818, + "step": 8670 + }, + { + "epoch": 0.3351480752152593, + "grad_norm": 2.3636281490325928, + "learning_rate": 0.00017765936908760956, + "loss": 0.3317, + "step": 8680 + }, + { + "epoch": 0.33553419050928607, + "grad_norm": 3.3569977283477783, + "learning_rate": 0.0001776336280680078, + "loss": 0.4388, + "step": 8690 + }, + { + "epoch": 0.3359203058033129, + "grad_norm": 1.2452306747436523, + "learning_rate": 0.00017760788704840598, + "loss": 0.1368, + "step": 8700 + }, + { + "epoch": 0.33630642109733966, + "grad_norm": 0.0380173958837986, + "learning_rate": 0.0001775821460288042, + "loss": 0.3264, + "step": 8710 + }, + { + "epoch": 0.3366925363913665, + "grad_norm": 1.5271002054214478, + "learning_rate": 0.00017755640500920244, + "loss": 0.2943, + "step": 8720 + }, + { + "epoch": 0.33707865168539325, + "grad_norm": 0.9701687693595886, + "learning_rate": 0.00017753066398960062, + "loss": 0.353, + "step": 8730 + }, + { + "epoch": 0.33746476697942007, + "grad_norm": 1.9296154975891113, + "learning_rate": 0.00017750492296999886, + "loss": 0.3776, + "step": 8740 + }, + { + "epoch": 0.33785088227344684, + "grad_norm": 1.2136276960372925, + "learning_rate": 0.00017747918195039705, + "loss": 0.5126, + "step": 8750 + }, + { + "epoch": 0.33823699756747366, + "grad_norm": 1.7323212623596191, + "learning_rate": 0.0001774534409307953, + "loss": 0.3477, + "step": 8760 + }, + { + "epoch": 0.33862311286150043, + "grad_norm": 1.164534091949463, + "learning_rate": 0.0001774276999111935, + "loss": 0.4053, + "step": 8770 + }, + { + "epoch": 0.33900922815552725, + "grad_norm": 0.42989471554756165, + "learning_rate": 0.0001774019588915917, + "loss": 0.3026, + "step": 8780 + }, + { + "epoch": 0.339395343449554, + "grad_norm": 2.357590436935425, + "learning_rate": 0.00017737621787198993, + "loss": 0.3869, + "step": 8790 + }, + { + "epoch": 0.33978145874358084, + "grad_norm": 1.9374550580978394, + "learning_rate": 0.00017735047685238812, + "loss": 0.2975, + "step": 8800 + }, + { + "epoch": 0.3401675740376076, + "grad_norm": 4.8107428550720215, + "learning_rate": 0.00017732473583278636, + "loss": 0.3959, + "step": 8810 + }, + { + "epoch": 0.34055368933163443, + "grad_norm": 1.938700556755066, + "learning_rate": 0.00017729899481318454, + "loss": 0.3726, + "step": 8820 + }, + { + "epoch": 0.3409398046256612, + "grad_norm": 3.147167682647705, + "learning_rate": 0.00017727325379358278, + "loss": 0.1828, + "step": 8830 + }, + { + "epoch": 0.341325919919688, + "grad_norm": 1.8921313285827637, + "learning_rate": 0.000177247512773981, + "loss": 0.2038, + "step": 8840 + }, + { + "epoch": 0.3417120352137148, + "grad_norm": 0.9098349213600159, + "learning_rate": 0.00017722177175437918, + "loss": 0.2853, + "step": 8850 + }, + { + "epoch": 0.3420981505077416, + "grad_norm": 2.4006853103637695, + "learning_rate": 0.00017719603073477742, + "loss": 0.6054, + "step": 8860 + }, + { + "epoch": 0.3424842658017684, + "grad_norm": 1.9303867816925049, + "learning_rate": 0.0001771702897151756, + "loss": 0.3507, + "step": 8870 + }, + { + "epoch": 0.3428703810957952, + "grad_norm": 0.49361029267311096, + "learning_rate": 0.00017714454869557385, + "loss": 0.4661, + "step": 8880 + }, + { + "epoch": 0.343256496389822, + "grad_norm": 2.542618751525879, + "learning_rate": 0.00017711880767597204, + "loss": 0.6924, + "step": 8890 + }, + { + "epoch": 0.3436426116838488, + "grad_norm": 0.5868918895721436, + "learning_rate": 0.00017709306665637028, + "loss": 0.4507, + "step": 8900 + }, + { + "epoch": 0.3440287269778756, + "grad_norm": 2.4685137271881104, + "learning_rate": 0.0001770673256367685, + "loss": 0.4538, + "step": 8910 + }, + { + "epoch": 0.3444148422719024, + "grad_norm": 2.6662702560424805, + "learning_rate": 0.00017704158461716668, + "loss": 0.6181, + "step": 8920 + }, + { + "epoch": 0.3448009575659292, + "grad_norm": 1.705103874206543, + "learning_rate": 0.00017701584359756492, + "loss": 0.481, + "step": 8930 + }, + { + "epoch": 0.345187072859956, + "grad_norm": 2.0710952281951904, + "learning_rate": 0.0001769901025779631, + "loss": 0.4357, + "step": 8940 + }, + { + "epoch": 0.3455731881539828, + "grad_norm": 3.487117290496826, + "learning_rate": 0.00017696436155836134, + "loss": 0.4572, + "step": 8950 + }, + { + "epoch": 0.34595930344800957, + "grad_norm": 3.03472900390625, + "learning_rate": 0.00017693862053875956, + "loss": 0.4437, + "step": 8960 + }, + { + "epoch": 0.3463454187420364, + "grad_norm": 1.310692548751831, + "learning_rate": 0.00017691287951915777, + "loss": 0.4218, + "step": 8970 + }, + { + "epoch": 0.34673153403606316, + "grad_norm": 4.131219387054443, + "learning_rate": 0.00017688713849955598, + "loss": 0.445, + "step": 8980 + }, + { + "epoch": 0.34711764933009, + "grad_norm": 1.4199285507202148, + "learning_rate": 0.00017686139747995417, + "loss": 0.3093, + "step": 8990 + }, + { + "epoch": 0.34750376462411675, + "grad_norm": 1.9338914155960083, + "learning_rate": 0.0001768356564603524, + "loss": 0.275, + "step": 9000 + }, + { + "epoch": 0.3478898799181436, + "grad_norm": 3.425877332687378, + "learning_rate": 0.0001768099154407506, + "loss": 0.4621, + "step": 9010 + }, + { + "epoch": 0.34827599521217034, + "grad_norm": 3.6350486278533936, + "learning_rate": 0.00017678417442114884, + "loss": 0.3313, + "step": 9020 + }, + { + "epoch": 0.34866211050619716, + "grad_norm": 3.339202880859375, + "learning_rate": 0.00017675843340154705, + "loss": 0.5958, + "step": 9030 + }, + { + "epoch": 0.34904822580022393, + "grad_norm": 4.148682117462158, + "learning_rate": 0.00017673269238194526, + "loss": 0.2384, + "step": 9040 + }, + { + "epoch": 0.34943434109425076, + "grad_norm": 0.9697182178497314, + "learning_rate": 0.00017670695136234348, + "loss": 0.3119, + "step": 9050 + }, + { + "epoch": 0.3498204563882775, + "grad_norm": 0.53201824426651, + "learning_rate": 0.0001766812103427417, + "loss": 0.4339, + "step": 9060 + }, + { + "epoch": 0.35020657168230435, + "grad_norm": 0.9727185368537903, + "learning_rate": 0.0001766554693231399, + "loss": 0.3289, + "step": 9070 + }, + { + "epoch": 0.3505926869763311, + "grad_norm": 4.32904052734375, + "learning_rate": 0.00017662972830353812, + "loss": 0.4673, + "step": 9080 + }, + { + "epoch": 0.35097880227035794, + "grad_norm": 2.511558771133423, + "learning_rate": 0.00017660398728393633, + "loss": 0.2257, + "step": 9090 + }, + { + "epoch": 0.3513649175643847, + "grad_norm": 1.8378714323043823, + "learning_rate": 0.00017657824626433454, + "loss": 0.3977, + "step": 9100 + }, + { + "epoch": 0.3517510328584115, + "grad_norm": 1.3297137022018433, + "learning_rate": 0.00017655250524473276, + "loss": 0.3541, + "step": 9110 + }, + { + "epoch": 0.3521371481524383, + "grad_norm": 3.253089666366577, + "learning_rate": 0.00017652676422513097, + "loss": 0.6326, + "step": 9120 + }, + { + "epoch": 0.3525232634464651, + "grad_norm": 0.9691923260688782, + "learning_rate": 0.00017650102320552918, + "loss": 0.2206, + "step": 9130 + }, + { + "epoch": 0.3529093787404919, + "grad_norm": 1.570204496383667, + "learning_rate": 0.0001764752821859274, + "loss": 0.2769, + "step": 9140 + }, + { + "epoch": 0.3532954940345187, + "grad_norm": 1.9307161569595337, + "learning_rate": 0.0001764495411663256, + "loss": 0.3149, + "step": 9150 + }, + { + "epoch": 0.3536816093285455, + "grad_norm": 2.783297300338745, + "learning_rate": 0.00017642380014672382, + "loss": 0.3912, + "step": 9160 + }, + { + "epoch": 0.3540677246225723, + "grad_norm": 2.193371057510376, + "learning_rate": 0.00017639805912712204, + "loss": 0.3782, + "step": 9170 + }, + { + "epoch": 0.3544538399165991, + "grad_norm": 2.3460335731506348, + "learning_rate": 0.00017637231810752025, + "loss": 0.5051, + "step": 9180 + }, + { + "epoch": 0.3548399552106259, + "grad_norm": 2.4668326377868652, + "learning_rate": 0.00017634657708791846, + "loss": 0.2899, + "step": 9190 + }, + { + "epoch": 0.3552260705046527, + "grad_norm": 2.004683017730713, + "learning_rate": 0.00017632083606831668, + "loss": 0.3137, + "step": 9200 + }, + { + "epoch": 0.3556121857986795, + "grad_norm": 6.333971977233887, + "learning_rate": 0.0001762950950487149, + "loss": 0.5027, + "step": 9210 + }, + { + "epoch": 0.3559983010927063, + "grad_norm": 1.7840352058410645, + "learning_rate": 0.0001762693540291131, + "loss": 0.3988, + "step": 9220 + }, + { + "epoch": 0.35638441638673307, + "grad_norm": 0.9257024526596069, + "learning_rate": 0.00017624361300951132, + "loss": 0.3662, + "step": 9230 + }, + { + "epoch": 0.3567705316807599, + "grad_norm": 2.582887887954712, + "learning_rate": 0.00017621787198990953, + "loss": 0.2863, + "step": 9240 + }, + { + "epoch": 0.35715664697478666, + "grad_norm": 3.119943380355835, + "learning_rate": 0.00017619213097030774, + "loss": 0.4041, + "step": 9250 + }, + { + "epoch": 0.3575427622688135, + "grad_norm": 2.2561371326446533, + "learning_rate": 0.00017616638995070596, + "loss": 0.3969, + "step": 9260 + }, + { + "epoch": 0.35792887756284025, + "grad_norm": 2.104891538619995, + "learning_rate": 0.00017614064893110417, + "loss": 0.3216, + "step": 9270 + }, + { + "epoch": 0.3583149928568671, + "grad_norm": 1.6922805309295654, + "learning_rate": 0.00017611490791150238, + "loss": 0.3828, + "step": 9280 + }, + { + "epoch": 0.35870110815089384, + "grad_norm": 1.0928469896316528, + "learning_rate": 0.0001760891668919006, + "loss": 0.3225, + "step": 9290 + }, + { + "epoch": 0.35908722344492067, + "grad_norm": 2.4089863300323486, + "learning_rate": 0.0001760634258722988, + "loss": 0.4143, + "step": 9300 + }, + { + "epoch": 0.35947333873894743, + "grad_norm": 0.5562119483947754, + "learning_rate": 0.00017603768485269702, + "loss": 0.4597, + "step": 9310 + }, + { + "epoch": 0.35985945403297426, + "grad_norm": 1.3904486894607544, + "learning_rate": 0.00017601194383309524, + "loss": 0.4462, + "step": 9320 + }, + { + "epoch": 0.360245569327001, + "grad_norm": 2.1393306255340576, + "learning_rate": 0.00017598620281349345, + "loss": 0.2613, + "step": 9330 + }, + { + "epoch": 0.36063168462102785, + "grad_norm": 1.3657029867172241, + "learning_rate": 0.00017596046179389166, + "loss": 0.4968, + "step": 9340 + }, + { + "epoch": 0.3610177999150546, + "grad_norm": 2.424880027770996, + "learning_rate": 0.00017593472077428988, + "loss": 0.5982, + "step": 9350 + }, + { + "epoch": 0.36140391520908144, + "grad_norm": 6.178807735443115, + "learning_rate": 0.0001759089797546881, + "loss": 0.5355, + "step": 9360 + }, + { + "epoch": 0.3617900305031082, + "grad_norm": 1.5572419166564941, + "learning_rate": 0.0001758832387350863, + "loss": 0.4435, + "step": 9370 + }, + { + "epoch": 0.36217614579713503, + "grad_norm": 0.46649104356765747, + "learning_rate": 0.00017585749771548452, + "loss": 0.352, + "step": 9380 + }, + { + "epoch": 0.3625622610911618, + "grad_norm": 1.9611142873764038, + "learning_rate": 0.00017583175669588276, + "loss": 0.2684, + "step": 9390 + }, + { + "epoch": 0.3629483763851886, + "grad_norm": 1.7648595571517944, + "learning_rate": 0.00017580601567628094, + "loss": 0.3186, + "step": 9400 + }, + { + "epoch": 0.3633344916792154, + "grad_norm": 1.7970843315124512, + "learning_rate": 0.00017578027465667916, + "loss": 0.5339, + "step": 9410 + }, + { + "epoch": 0.3637206069732422, + "grad_norm": 3.084897994995117, + "learning_rate": 0.00017575453363707737, + "loss": 0.5143, + "step": 9420 + }, + { + "epoch": 0.36410672226726903, + "grad_norm": 1.440626621246338, + "learning_rate": 0.00017572879261747558, + "loss": 0.4067, + "step": 9430 + }, + { + "epoch": 0.3644928375612958, + "grad_norm": 0.44918450713157654, + "learning_rate": 0.0001757030515978738, + "loss": 0.2306, + "step": 9440 + }, + { + "epoch": 0.3648789528553226, + "grad_norm": 2.617272138595581, + "learning_rate": 0.000175677310578272, + "loss": 0.3166, + "step": 9450 + }, + { + "epoch": 0.3652650681493494, + "grad_norm": 2.575073719024658, + "learning_rate": 0.00017565156955867025, + "loss": 0.6645, + "step": 9460 + }, + { + "epoch": 0.3656511834433762, + "grad_norm": 0.9430664777755737, + "learning_rate": 0.00017562582853906844, + "loss": 0.2753, + "step": 9470 + }, + { + "epoch": 0.366037298737403, + "grad_norm": 1.9400445222854614, + "learning_rate": 0.00017560008751946665, + "loss": 0.4689, + "step": 9480 + }, + { + "epoch": 0.3664234140314298, + "grad_norm": 4.0443220138549805, + "learning_rate": 0.00017557434649986486, + "loss": 0.5373, + "step": 9490 + }, + { + "epoch": 0.3668095293254566, + "grad_norm": 3.4999184608459473, + "learning_rate": 0.00017554860548026308, + "loss": 0.3412, + "step": 9500 + }, + { + "epoch": 0.3671956446194834, + "grad_norm": 2.2023515701293945, + "learning_rate": 0.0001755228644606613, + "loss": 0.3385, + "step": 9510 + }, + { + "epoch": 0.36758175991351016, + "grad_norm": 1.213641881942749, + "learning_rate": 0.0001754971234410595, + "loss": 0.4785, + "step": 9520 + }, + { + "epoch": 0.367967875207537, + "grad_norm": 0.4178420603275299, + "learning_rate": 0.00017547138242145774, + "loss": 0.2605, + "step": 9530 + }, + { + "epoch": 0.36835399050156376, + "grad_norm": 2.676564931869507, + "learning_rate": 0.00017544564140185593, + "loss": 0.5297, + "step": 9540 + }, + { + "epoch": 0.3687401057955906, + "grad_norm": 0.8604353070259094, + "learning_rate": 0.00017541990038225414, + "loss": 0.3983, + "step": 9550 + }, + { + "epoch": 0.36912622108961735, + "grad_norm": 1.298893690109253, + "learning_rate": 0.00017539415936265236, + "loss": 0.3229, + "step": 9560 + }, + { + "epoch": 0.36951233638364417, + "grad_norm": 4.109025478363037, + "learning_rate": 0.00017536841834305057, + "loss": 0.519, + "step": 9570 + }, + { + "epoch": 0.36989845167767094, + "grad_norm": 3.440915584564209, + "learning_rate": 0.0001753426773234488, + "loss": 0.4061, + "step": 9580 + }, + { + "epoch": 0.37028456697169776, + "grad_norm": 0.1484186202287674, + "learning_rate": 0.000175316936303847, + "loss": 0.3443, + "step": 9590 + }, + { + "epoch": 0.3706706822657245, + "grad_norm": 3.114328145980835, + "learning_rate": 0.00017529119528424524, + "loss": 0.2451, + "step": 9600 + }, + { + "epoch": 0.37105679755975135, + "grad_norm": 1.8218796253204346, + "learning_rate": 0.00017526545426464342, + "loss": 0.2511, + "step": 9610 + }, + { + "epoch": 0.3714429128537781, + "grad_norm": 1.0732795000076294, + "learning_rate": 0.00017523971324504164, + "loss": 0.1581, + "step": 9620 + }, + { + "epoch": 0.37182902814780494, + "grad_norm": 1.0567959547042847, + "learning_rate": 0.00017521397222543985, + "loss": 0.1924, + "step": 9630 + }, + { + "epoch": 0.3722151434418317, + "grad_norm": 0.3467637896537781, + "learning_rate": 0.00017518823120583806, + "loss": 0.3571, + "step": 9640 + }, + { + "epoch": 0.37260125873585853, + "grad_norm": 2.6293838024139404, + "learning_rate": 0.0001751624901862363, + "loss": 0.3282, + "step": 9650 + }, + { + "epoch": 0.3729873740298853, + "grad_norm": 1.159696102142334, + "learning_rate": 0.0001751367491666345, + "loss": 0.2636, + "step": 9660 + }, + { + "epoch": 0.3733734893239121, + "grad_norm": 0.6884826421737671, + "learning_rate": 0.00017511100814703273, + "loss": 0.2842, + "step": 9670 + }, + { + "epoch": 0.3737596046179389, + "grad_norm": 3.789825201034546, + "learning_rate": 0.00017508526712743091, + "loss": 0.599, + "step": 9680 + }, + { + "epoch": 0.3741457199119657, + "grad_norm": 1.0705493688583374, + "learning_rate": 0.00017505952610782913, + "loss": 0.1746, + "step": 9690 + }, + { + "epoch": 0.37453183520599254, + "grad_norm": 1.8735803365707397, + "learning_rate": 0.00017503378508822734, + "loss": 0.3259, + "step": 9700 + }, + { + "epoch": 0.3749179505000193, + "grad_norm": 1.2987112998962402, + "learning_rate": 0.00017500804406862555, + "loss": 0.5738, + "step": 9710 + }, + { + "epoch": 0.3753040657940461, + "grad_norm": 1.5362507104873657, + "learning_rate": 0.0001749823030490238, + "loss": 0.3815, + "step": 9720 + }, + { + "epoch": 0.3756901810880729, + "grad_norm": 0.1640123724937439, + "learning_rate": 0.00017495656202942198, + "loss": 0.3672, + "step": 9730 + }, + { + "epoch": 0.3760762963820997, + "grad_norm": 0.6714594960212708, + "learning_rate": 0.00017493082100982022, + "loss": 0.2849, + "step": 9740 + }, + { + "epoch": 0.3764624116761265, + "grad_norm": 4.330246448516846, + "learning_rate": 0.0001749050799902184, + "loss": 0.4023, + "step": 9750 + }, + { + "epoch": 0.3768485269701533, + "grad_norm": 0.8616659641265869, + "learning_rate": 0.00017487933897061665, + "loss": 0.4434, + "step": 9760 + }, + { + "epoch": 0.3772346422641801, + "grad_norm": 2.6581578254699707, + "learning_rate": 0.00017485359795101486, + "loss": 0.4854, + "step": 9770 + }, + { + "epoch": 0.3776207575582069, + "grad_norm": 1.8269850015640259, + "learning_rate": 0.00017482785693141305, + "loss": 0.6033, + "step": 9780 + }, + { + "epoch": 0.37800687285223367, + "grad_norm": 2.256073236465454, + "learning_rate": 0.0001748021159118113, + "loss": 0.5317, + "step": 9790 + }, + { + "epoch": 0.3783929881462605, + "grad_norm": 0.8793076872825623, + "learning_rate": 0.00017477637489220947, + "loss": 0.3883, + "step": 9800 + }, + { + "epoch": 0.37877910344028726, + "grad_norm": 1.71831214427948, + "learning_rate": 0.00017475063387260772, + "loss": 0.2473, + "step": 9810 + }, + { + "epoch": 0.3791652187343141, + "grad_norm": 3.4802069664001465, + "learning_rate": 0.0001747248928530059, + "loss": 0.4847, + "step": 9820 + }, + { + "epoch": 0.37955133402834085, + "grad_norm": 5.419053077697754, + "learning_rate": 0.00017469915183340414, + "loss": 0.3668, + "step": 9830 + }, + { + "epoch": 0.37993744932236767, + "grad_norm": 1.567060112953186, + "learning_rate": 0.00017467341081380236, + "loss": 0.3342, + "step": 9840 + }, + { + "epoch": 0.38032356461639444, + "grad_norm": 3.0100274085998535, + "learning_rate": 0.00017464766979420054, + "loss": 0.476, + "step": 9850 + }, + { + "epoch": 0.38070967991042126, + "grad_norm": 0.7659344673156738, + "learning_rate": 0.00017462192877459878, + "loss": 0.2608, + "step": 9860 + }, + { + "epoch": 0.38109579520444803, + "grad_norm": 3.9540984630584717, + "learning_rate": 0.00017459618775499697, + "loss": 0.763, + "step": 9870 + }, + { + "epoch": 0.38148191049847485, + "grad_norm": 0.8768689036369324, + "learning_rate": 0.0001745704467353952, + "loss": 0.3365, + "step": 9880 + }, + { + "epoch": 0.3818680257925016, + "grad_norm": 0.9985928535461426, + "learning_rate": 0.00017454470571579342, + "loss": 0.3116, + "step": 9890 + }, + { + "epoch": 0.38225414108652844, + "grad_norm": 2.0326671600341797, + "learning_rate": 0.00017451896469619163, + "loss": 0.289, + "step": 9900 + }, + { + "epoch": 0.3826402563805552, + "grad_norm": 3.2696290016174316, + "learning_rate": 0.00017449322367658985, + "loss": 0.4097, + "step": 9910 + }, + { + "epoch": 0.38302637167458203, + "grad_norm": 3.048860788345337, + "learning_rate": 0.00017446748265698803, + "loss": 0.5181, + "step": 9920 + }, + { + "epoch": 0.3834124869686088, + "grad_norm": 1.7899913787841797, + "learning_rate": 0.00017444174163738627, + "loss": 0.2166, + "step": 9930 + }, + { + "epoch": 0.3837986022626356, + "grad_norm": 3.6762959957122803, + "learning_rate": 0.00017441600061778446, + "loss": 0.4971, + "step": 9940 + }, + { + "epoch": 0.3841847175566624, + "grad_norm": 0.9108519554138184, + "learning_rate": 0.0001743902595981827, + "loss": 0.4974, + "step": 9950 + }, + { + "epoch": 0.3845708328506892, + "grad_norm": 4.062527656555176, + "learning_rate": 0.00017436451857858091, + "loss": 0.4448, + "step": 9960 + }, + { + "epoch": 0.38495694814471604, + "grad_norm": 3.230902671813965, + "learning_rate": 0.00017433877755897913, + "loss": 0.2977, + "step": 9970 + }, + { + "epoch": 0.3853430634387428, + "grad_norm": 3.8190758228302, + "learning_rate": 0.00017431303653937734, + "loss": 0.4887, + "step": 9980 + }, + { + "epoch": 0.38572917873276963, + "grad_norm": 0.9079695343971252, + "learning_rate": 0.00017428729551977553, + "loss": 0.271, + "step": 9990 + }, + { + "epoch": 0.3861152940267964, + "grad_norm": 3.3730807304382324, + "learning_rate": 0.00017426155450017377, + "loss": 0.3782, + "step": 10000 + }, + { + "epoch": 0.3865014093208232, + "grad_norm": 1.07533860206604, + "learning_rate": 0.00017423581348057195, + "loss": 0.3905, + "step": 10010 + }, + { + "epoch": 0.38688752461485, + "grad_norm": 1.3856415748596191, + "learning_rate": 0.0001742100724609702, + "loss": 0.3757, + "step": 10020 + }, + { + "epoch": 0.3872736399088768, + "grad_norm": 5.751671314239502, + "learning_rate": 0.0001741843314413684, + "loss": 0.6657, + "step": 10030 + }, + { + "epoch": 0.3876597552029036, + "grad_norm": 0.6837680339813232, + "learning_rate": 0.00017415859042176662, + "loss": 0.2318, + "step": 10040 + }, + { + "epoch": 0.3880458704969304, + "grad_norm": 2.770787239074707, + "learning_rate": 0.00017413284940216483, + "loss": 0.3706, + "step": 10050 + }, + { + "epoch": 0.38843198579095717, + "grad_norm": 2.3058855533599854, + "learning_rate": 0.00017410710838256302, + "loss": 0.1641, + "step": 10060 + }, + { + "epoch": 0.388818101084984, + "grad_norm": 1.894718885421753, + "learning_rate": 0.00017408136736296126, + "loss": 0.4752, + "step": 10070 + }, + { + "epoch": 0.38920421637901076, + "grad_norm": 1.8346868753433228, + "learning_rate": 0.00017405562634335947, + "loss": 0.5007, + "step": 10080 + }, + { + "epoch": 0.3895903316730376, + "grad_norm": 5.277680397033691, + "learning_rate": 0.0001740298853237577, + "loss": 0.4399, + "step": 10090 + }, + { + "epoch": 0.38997644696706435, + "grad_norm": 1.306093692779541, + "learning_rate": 0.0001740041443041559, + "loss": 0.371, + "step": 10100 + }, + { + "epoch": 0.3903625622610912, + "grad_norm": 3.0306456089019775, + "learning_rate": 0.00017397840328455411, + "loss": 0.2515, + "step": 10110 + }, + { + "epoch": 0.39074867755511794, + "grad_norm": 0.7951543927192688, + "learning_rate": 0.00017395266226495233, + "loss": 0.3775, + "step": 10120 + }, + { + "epoch": 0.39113479284914476, + "grad_norm": 5.185150146484375, + "learning_rate": 0.00017392692124535051, + "loss": 0.3591, + "step": 10130 + }, + { + "epoch": 0.39152090814317153, + "grad_norm": 1.1718593835830688, + "learning_rate": 0.00017390118022574875, + "loss": 0.5484, + "step": 10140 + }, + { + "epoch": 0.39190702343719835, + "grad_norm": 1.6352128982543945, + "learning_rate": 0.00017387543920614697, + "loss": 0.2817, + "step": 10150 + }, + { + "epoch": 0.3922931387312251, + "grad_norm": 2.4863786697387695, + "learning_rate": 0.00017384969818654518, + "loss": 0.4027, + "step": 10160 + }, + { + "epoch": 0.39267925402525194, + "grad_norm": 2.069805383682251, + "learning_rate": 0.0001738239571669434, + "loss": 0.3559, + "step": 10170 + }, + { + "epoch": 0.3930653693192787, + "grad_norm": 1.671980619430542, + "learning_rate": 0.0001737982161473416, + "loss": 0.4405, + "step": 10180 + }, + { + "epoch": 0.39345148461330554, + "grad_norm": 4.298947334289551, + "learning_rate": 0.00017377247512773982, + "loss": 0.3005, + "step": 10190 + }, + { + "epoch": 0.3938375999073323, + "grad_norm": 0.4142851233482361, + "learning_rate": 0.000173746734108138, + "loss": 0.4248, + "step": 10200 + }, + { + "epoch": 0.3942237152013591, + "grad_norm": 3.5962865352630615, + "learning_rate": 0.00017372099308853625, + "loss": 0.27, + "step": 10210 + }, + { + "epoch": 0.39460983049538595, + "grad_norm": 2.20154070854187, + "learning_rate": 0.00017369525206893446, + "loss": 0.2858, + "step": 10220 + }, + { + "epoch": 0.3949959457894127, + "grad_norm": 0.2400553673505783, + "learning_rate": 0.00017366951104933267, + "loss": 0.2806, + "step": 10230 + }, + { + "epoch": 0.39538206108343954, + "grad_norm": 1.817741870880127, + "learning_rate": 0.0001736437700297309, + "loss": 0.3647, + "step": 10240 + }, + { + "epoch": 0.3957681763774663, + "grad_norm": 4.890044689178467, + "learning_rate": 0.0001736180290101291, + "loss": 0.4435, + "step": 10250 + }, + { + "epoch": 0.39615429167149313, + "grad_norm": 0.3407624065876007, + "learning_rate": 0.00017359228799052731, + "loss": 0.4857, + "step": 10260 + }, + { + "epoch": 0.3965404069655199, + "grad_norm": 2.4883463382720947, + "learning_rate": 0.00017356654697092553, + "loss": 0.2667, + "step": 10270 + }, + { + "epoch": 0.3969265222595467, + "grad_norm": 2.343823194503784, + "learning_rate": 0.00017354080595132374, + "loss": 0.3711, + "step": 10280 + }, + { + "epoch": 0.3973126375535735, + "grad_norm": 0.2056214064359665, + "learning_rate": 0.00017351506493172195, + "loss": 0.2695, + "step": 10290 + }, + { + "epoch": 0.3976987528476003, + "grad_norm": 0.20321065187454224, + "learning_rate": 0.00017348932391212017, + "loss": 0.3079, + "step": 10300 + }, + { + "epoch": 0.3980848681416271, + "grad_norm": 0.7993821501731873, + "learning_rate": 0.00017346358289251838, + "loss": 0.3599, + "step": 10310 + }, + { + "epoch": 0.3984709834356539, + "grad_norm": 2.0987348556518555, + "learning_rate": 0.0001734378418729166, + "loss": 0.3259, + "step": 10320 + }, + { + "epoch": 0.39885709872968067, + "grad_norm": 2.474246025085449, + "learning_rate": 0.0001734121008533148, + "loss": 0.3398, + "step": 10330 + }, + { + "epoch": 0.3992432140237075, + "grad_norm": 2.341064214706421, + "learning_rate": 0.00017338635983371302, + "loss": 0.5264, + "step": 10340 + }, + { + "epoch": 0.39962932931773426, + "grad_norm": 1.587437629699707, + "learning_rate": 0.00017336061881411123, + "loss": 0.4228, + "step": 10350 + }, + { + "epoch": 0.4000154446117611, + "grad_norm": 0.6692029237747192, + "learning_rate": 0.00017333487779450945, + "loss": 0.3576, + "step": 10360 + }, + { + "epoch": 0.40040155990578785, + "grad_norm": 2.088212251663208, + "learning_rate": 0.00017330913677490766, + "loss": 0.3096, + "step": 10370 + }, + { + "epoch": 0.4007876751998147, + "grad_norm": 1.5051954984664917, + "learning_rate": 0.00017328339575530587, + "loss": 0.3753, + "step": 10380 + }, + { + "epoch": 0.40117379049384144, + "grad_norm": 2.02595591545105, + "learning_rate": 0.0001732576547357041, + "loss": 0.3339, + "step": 10390 + }, + { + "epoch": 0.40155990578786827, + "grad_norm": 1.3062909841537476, + "learning_rate": 0.0001732319137161023, + "loss": 0.4301, + "step": 10400 + }, + { + "epoch": 0.40194602108189503, + "grad_norm": 2.5890421867370605, + "learning_rate": 0.00017320617269650051, + "loss": 0.3047, + "step": 10410 + }, + { + "epoch": 0.40233213637592186, + "grad_norm": 1.5994844436645508, + "learning_rate": 0.00017318043167689873, + "loss": 0.4158, + "step": 10420 + }, + { + "epoch": 0.4027182516699486, + "grad_norm": 0.5470211505889893, + "learning_rate": 0.00017315469065729694, + "loss": 0.4513, + "step": 10430 + }, + { + "epoch": 0.40310436696397545, + "grad_norm": 2.216935634613037, + "learning_rate": 0.00017312894963769515, + "loss": 0.5123, + "step": 10440 + }, + { + "epoch": 0.4034904822580022, + "grad_norm": 2.354724645614624, + "learning_rate": 0.00017310320861809337, + "loss": 0.2804, + "step": 10450 + }, + { + "epoch": 0.40387659755202904, + "grad_norm": 4.514159202575684, + "learning_rate": 0.00017307746759849158, + "loss": 0.3317, + "step": 10460 + }, + { + "epoch": 0.4042627128460558, + "grad_norm": 0.9874318242073059, + "learning_rate": 0.0001730517265788898, + "loss": 0.1948, + "step": 10470 + }, + { + "epoch": 0.40464882814008263, + "grad_norm": 2.0725696086883545, + "learning_rate": 0.000173025985559288, + "loss": 0.3627, + "step": 10480 + }, + { + "epoch": 0.40503494343410945, + "grad_norm": 2.4061577320098877, + "learning_rate": 0.00017300024453968622, + "loss": 0.3074, + "step": 10490 + }, + { + "epoch": 0.4054210587281362, + "grad_norm": 1.3369660377502441, + "learning_rate": 0.00017297450352008443, + "loss": 0.533, + "step": 10500 + }, + { + "epoch": 0.40580717402216304, + "grad_norm": 1.2730306386947632, + "learning_rate": 0.00017294876250048265, + "loss": 0.4688, + "step": 10510 + }, + { + "epoch": 0.4061932893161898, + "grad_norm": 0.6753021478652954, + "learning_rate": 0.00017292302148088086, + "loss": 0.4427, + "step": 10520 + }, + { + "epoch": 0.40657940461021663, + "grad_norm": 1.7279945611953735, + "learning_rate": 0.0001728972804612791, + "loss": 0.4921, + "step": 10530 + }, + { + "epoch": 0.4069655199042434, + "grad_norm": 0.9288708567619324, + "learning_rate": 0.0001728715394416773, + "loss": 0.363, + "step": 10540 + }, + { + "epoch": 0.4073516351982702, + "grad_norm": 0.5325084924697876, + "learning_rate": 0.0001728457984220755, + "loss": 0.4095, + "step": 10550 + }, + { + "epoch": 0.407737750492297, + "grad_norm": 1.2030489444732666, + "learning_rate": 0.0001728200574024737, + "loss": 0.3499, + "step": 10560 + }, + { + "epoch": 0.4081238657863238, + "grad_norm": 3.8157269954681396, + "learning_rate": 0.00017279431638287193, + "loss": 0.1622, + "step": 10570 + }, + { + "epoch": 0.4085099810803506, + "grad_norm": 0.6373336911201477, + "learning_rate": 0.00017276857536327017, + "loss": 0.4657, + "step": 10580 + }, + { + "epoch": 0.4088960963743774, + "grad_norm": 2.2850074768066406, + "learning_rate": 0.00017274283434366835, + "loss": 0.3585, + "step": 10590 + }, + { + "epoch": 0.40928221166840417, + "grad_norm": 0.8831659555435181, + "learning_rate": 0.0001727170933240666, + "loss": 0.293, + "step": 10600 + }, + { + "epoch": 0.409668326962431, + "grad_norm": 5.1165995597839355, + "learning_rate": 0.00017269135230446478, + "loss": 0.6539, + "step": 10610 + }, + { + "epoch": 0.41005444225645776, + "grad_norm": 4.901204586029053, + "learning_rate": 0.000172665611284863, + "loss": 0.4628, + "step": 10620 + }, + { + "epoch": 0.4104405575504846, + "grad_norm": 2.1492419242858887, + "learning_rate": 0.0001726398702652612, + "loss": 0.277, + "step": 10630 + }, + { + "epoch": 0.41082667284451135, + "grad_norm": 3.56510853767395, + "learning_rate": 0.00017261412924565942, + "loss": 0.4696, + "step": 10640 + }, + { + "epoch": 0.4112127881385382, + "grad_norm": 2.054769992828369, + "learning_rate": 0.00017258838822605766, + "loss": 0.4093, + "step": 10650 + }, + { + "epoch": 0.41159890343256494, + "grad_norm": 2.133474826812744, + "learning_rate": 0.00017256264720645585, + "loss": 0.3604, + "step": 10660 + }, + { + "epoch": 0.41198501872659177, + "grad_norm": 2.5062367916107178, + "learning_rate": 0.0001725369061868541, + "loss": 0.3916, + "step": 10670 + }, + { + "epoch": 0.41237113402061853, + "grad_norm": 0.431570827960968, + "learning_rate": 0.00017251116516725227, + "loss": 0.4048, + "step": 10680 + }, + { + "epoch": 0.41275724931464536, + "grad_norm": 1.2092580795288086, + "learning_rate": 0.0001724854241476505, + "loss": 0.602, + "step": 10690 + }, + { + "epoch": 0.4131433646086721, + "grad_norm": 2.712398052215576, + "learning_rate": 0.00017245968312804873, + "loss": 0.4172, + "step": 10700 + }, + { + "epoch": 0.41352947990269895, + "grad_norm": 3.914670467376709, + "learning_rate": 0.0001724339421084469, + "loss": 0.3843, + "step": 10710 + }, + { + "epoch": 0.4139155951967257, + "grad_norm": 1.7062132358551025, + "learning_rate": 0.00017240820108884515, + "loss": 0.343, + "step": 10720 + }, + { + "epoch": 0.41430171049075254, + "grad_norm": 0.5837095379829407, + "learning_rate": 0.00017238246006924334, + "loss": 0.3872, + "step": 10730 + }, + { + "epoch": 0.41468782578477936, + "grad_norm": 1.098900556564331, + "learning_rate": 0.00017235671904964158, + "loss": 0.2062, + "step": 10740 + }, + { + "epoch": 0.41507394107880613, + "grad_norm": 1.2533438205718994, + "learning_rate": 0.00017233097803003977, + "loss": 0.141, + "step": 10750 + }, + { + "epoch": 0.41546005637283295, + "grad_norm": 0.8688085079193115, + "learning_rate": 0.00017230523701043798, + "loss": 0.3686, + "step": 10760 + }, + { + "epoch": 0.4158461716668597, + "grad_norm": 1.868402361869812, + "learning_rate": 0.00017227949599083622, + "loss": 0.449, + "step": 10770 + }, + { + "epoch": 0.41623228696088654, + "grad_norm": 0.7168850898742676, + "learning_rate": 0.0001722537549712344, + "loss": 0.2317, + "step": 10780 + }, + { + "epoch": 0.4166184022549133, + "grad_norm": 3.1062309741973877, + "learning_rate": 0.00017222801395163265, + "loss": 0.4655, + "step": 10790 + }, + { + "epoch": 0.41700451754894013, + "grad_norm": 2.7296605110168457, + "learning_rate": 0.00017220227293203083, + "loss": 0.3934, + "step": 10800 + }, + { + "epoch": 0.4173906328429669, + "grad_norm": 2.3148224353790283, + "learning_rate": 0.00017217653191242907, + "loss": 0.2367, + "step": 10810 + }, + { + "epoch": 0.4177767481369937, + "grad_norm": 0.7049677968025208, + "learning_rate": 0.00017215079089282726, + "loss": 0.3157, + "step": 10820 + }, + { + "epoch": 0.4181628634310205, + "grad_norm": 3.3960344791412354, + "learning_rate": 0.00017212504987322547, + "loss": 0.4945, + "step": 10830 + }, + { + "epoch": 0.4185489787250473, + "grad_norm": 2.606316566467285, + "learning_rate": 0.0001720993088536237, + "loss": 0.4056, + "step": 10840 + }, + { + "epoch": 0.4189350940190741, + "grad_norm": 1.7469319105148315, + "learning_rate": 0.0001720735678340219, + "loss": 0.4176, + "step": 10850 + }, + { + "epoch": 0.4193212093131009, + "grad_norm": 0.8538552522659302, + "learning_rate": 0.00017204782681442014, + "loss": 0.3025, + "step": 10860 + }, + { + "epoch": 0.4197073246071277, + "grad_norm": 1.9576159715652466, + "learning_rate": 0.00017202208579481833, + "loss": 0.5626, + "step": 10870 + }, + { + "epoch": 0.4200934399011545, + "grad_norm": 0.8435356616973877, + "learning_rate": 0.00017199634477521657, + "loss": 0.2397, + "step": 10880 + }, + { + "epoch": 0.42047955519518126, + "grad_norm": 1.3026552200317383, + "learning_rate": 0.00017197060375561478, + "loss": 0.4793, + "step": 10890 + }, + { + "epoch": 0.4208656704892081, + "grad_norm": 1.8935116529464722, + "learning_rate": 0.00017194486273601297, + "loss": 0.2459, + "step": 10900 + }, + { + "epoch": 0.42125178578323486, + "grad_norm": 0.7297415137290955, + "learning_rate": 0.0001719191217164112, + "loss": 0.4115, + "step": 10910 + }, + { + "epoch": 0.4216379010772617, + "grad_norm": 2.730445146560669, + "learning_rate": 0.0001718933806968094, + "loss": 0.3467, + "step": 10920 + }, + { + "epoch": 0.42202401637128845, + "grad_norm": 1.5462249517440796, + "learning_rate": 0.00017186763967720763, + "loss": 0.2319, + "step": 10930 + }, + { + "epoch": 0.42241013166531527, + "grad_norm": 2.173388957977295, + "learning_rate": 0.00017184189865760582, + "loss": 0.3664, + "step": 10940 + }, + { + "epoch": 0.42279624695934204, + "grad_norm": 0.9086957573890686, + "learning_rate": 0.00017181615763800406, + "loss": 0.3928, + "step": 10950 + }, + { + "epoch": 0.42318236225336886, + "grad_norm": 1.6344754695892334, + "learning_rate": 0.00017179041661840227, + "loss": 0.32, + "step": 10960 + }, + { + "epoch": 0.4235684775473956, + "grad_norm": 3.7620887756347656, + "learning_rate": 0.00017176467559880049, + "loss": 0.3998, + "step": 10970 + }, + { + "epoch": 0.42395459284142245, + "grad_norm": 2.3914058208465576, + "learning_rate": 0.0001717389345791987, + "loss": 0.3003, + "step": 10980 + }, + { + "epoch": 0.4243407081354492, + "grad_norm": 1.1183325052261353, + "learning_rate": 0.00017171319355959689, + "loss": 0.2408, + "step": 10990 + }, + { + "epoch": 0.42472682342947604, + "grad_norm": 1.5570834875106812, + "learning_rate": 0.00017168745253999513, + "loss": 0.2638, + "step": 11000 + }, + { + "epoch": 0.42511293872350286, + "grad_norm": 1.4825866222381592, + "learning_rate": 0.0001716617115203933, + "loss": 0.2774, + "step": 11010 + }, + { + "epoch": 0.42549905401752963, + "grad_norm": 1.5424071550369263, + "learning_rate": 0.00017163597050079155, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.42588516931155646, + "grad_norm": 3.9182989597320557, + "learning_rate": 0.00017161022948118977, + "loss": 0.5319, + "step": 11030 + }, + { + "epoch": 0.4262712846055832, + "grad_norm": 3.7870359420776367, + "learning_rate": 0.00017158448846158798, + "loss": 0.316, + "step": 11040 + }, + { + "epoch": 0.42665739989961005, + "grad_norm": 3.6943869590759277, + "learning_rate": 0.0001715587474419862, + "loss": 0.4336, + "step": 11050 + }, + { + "epoch": 0.4270435151936368, + "grad_norm": 1.1681898832321167, + "learning_rate": 0.00017153300642238438, + "loss": 0.328, + "step": 11060 + }, + { + "epoch": 0.42742963048766364, + "grad_norm": 3.6428277492523193, + "learning_rate": 0.00017150726540278262, + "loss": 0.2609, + "step": 11070 + }, + { + "epoch": 0.4278157457816904, + "grad_norm": 0.8946434259414673, + "learning_rate": 0.00017148152438318083, + "loss": 0.5445, + "step": 11080 + }, + { + "epoch": 0.4282018610757172, + "grad_norm": 1.8038333654403687, + "learning_rate": 0.00017145578336357905, + "loss": 0.3548, + "step": 11090 + }, + { + "epoch": 0.428587976369744, + "grad_norm": 2.3430778980255127, + "learning_rate": 0.00017143004234397726, + "loss": 0.4831, + "step": 11100 + }, + { + "epoch": 0.4289740916637708, + "grad_norm": 1.3243132829666138, + "learning_rate": 0.00017140430132437547, + "loss": 0.5759, + "step": 11110 + }, + { + "epoch": 0.4293602069577976, + "grad_norm": 1.2575668096542358, + "learning_rate": 0.00017137856030477369, + "loss": 0.3289, + "step": 11120 + }, + { + "epoch": 0.4297463222518244, + "grad_norm": 1.5581884384155273, + "learning_rate": 0.00017135281928517187, + "loss": 0.309, + "step": 11130 + }, + { + "epoch": 0.4301324375458512, + "grad_norm": 2.736063241958618, + "learning_rate": 0.0001713270782655701, + "loss": 0.3964, + "step": 11140 + }, + { + "epoch": 0.430518552839878, + "grad_norm": 1.3930561542510986, + "learning_rate": 0.00017130133724596833, + "loss": 0.2277, + "step": 11150 + }, + { + "epoch": 0.43090466813390477, + "grad_norm": 1.7932826280593872, + "learning_rate": 0.00017127559622636654, + "loss": 0.2907, + "step": 11160 + }, + { + "epoch": 0.4312907834279316, + "grad_norm": 1.7017295360565186, + "learning_rate": 0.00017124985520676475, + "loss": 0.2986, + "step": 11170 + }, + { + "epoch": 0.43167689872195836, + "grad_norm": 1.3543587923049927, + "learning_rate": 0.00017122411418716297, + "loss": 0.4872, + "step": 11180 + }, + { + "epoch": 0.4320630140159852, + "grad_norm": 2.3927829265594482, + "learning_rate": 0.00017119837316756118, + "loss": 0.5281, + "step": 11190 + }, + { + "epoch": 0.43244912931001195, + "grad_norm": 2.916257619857788, + "learning_rate": 0.0001711726321479594, + "loss": 0.5364, + "step": 11200 + }, + { + "epoch": 0.43283524460403877, + "grad_norm": 4.717250823974609, + "learning_rate": 0.0001711468911283576, + "loss": 0.3744, + "step": 11210 + }, + { + "epoch": 0.43322135989806554, + "grad_norm": 1.806577444076538, + "learning_rate": 0.00017112115010875582, + "loss": 0.3377, + "step": 11220 + }, + { + "epoch": 0.43360747519209236, + "grad_norm": 3.186603546142578, + "learning_rate": 0.00017109540908915403, + "loss": 0.3421, + "step": 11230 + }, + { + "epoch": 0.43399359048611913, + "grad_norm": 5.388319969177246, + "learning_rate": 0.00017106966806955225, + "loss": 0.3589, + "step": 11240 + }, + { + "epoch": 0.43437970578014595, + "grad_norm": 1.368312954902649, + "learning_rate": 0.00017104392704995046, + "loss": 0.2677, + "step": 11250 + }, + { + "epoch": 0.4347658210741727, + "grad_norm": 0.9010117053985596, + "learning_rate": 0.00017101818603034867, + "loss": 0.3411, + "step": 11260 + }, + { + "epoch": 0.43515193636819954, + "grad_norm": 0.46370139718055725, + "learning_rate": 0.00017099244501074689, + "loss": 0.3531, + "step": 11270 + }, + { + "epoch": 0.43553805166222637, + "grad_norm": 2.778857469558716, + "learning_rate": 0.0001709667039911451, + "loss": 0.3953, + "step": 11280 + }, + { + "epoch": 0.43592416695625313, + "grad_norm": 0.45829036831855774, + "learning_rate": 0.0001709409629715433, + "loss": 0.3117, + "step": 11290 + }, + { + "epoch": 0.43631028225027996, + "grad_norm": 2.2053589820861816, + "learning_rate": 0.00017091522195194153, + "loss": 0.5104, + "step": 11300 + }, + { + "epoch": 0.4366963975443067, + "grad_norm": 5.166933059692383, + "learning_rate": 0.00017088948093233974, + "loss": 0.6913, + "step": 11310 + }, + { + "epoch": 0.43708251283833355, + "grad_norm": 1.5593189001083374, + "learning_rate": 0.00017086373991273795, + "loss": 0.305, + "step": 11320 + }, + { + "epoch": 0.4374686281323603, + "grad_norm": 2.01481556892395, + "learning_rate": 0.00017083799889313617, + "loss": 0.4576, + "step": 11330 + }, + { + "epoch": 0.43785474342638714, + "grad_norm": 2.200463056564331, + "learning_rate": 0.00017081225787353438, + "loss": 0.6491, + "step": 11340 + }, + { + "epoch": 0.4382408587204139, + "grad_norm": 0.8904009461402893, + "learning_rate": 0.0001707865168539326, + "loss": 0.134, + "step": 11350 + }, + { + "epoch": 0.43862697401444073, + "grad_norm": 0.7481307983398438, + "learning_rate": 0.0001707607758343308, + "loss": 0.5552, + "step": 11360 + }, + { + "epoch": 0.4390130893084675, + "grad_norm": 1.0893138647079468, + "learning_rate": 0.00017073503481472902, + "loss": 0.2369, + "step": 11370 + }, + { + "epoch": 0.4393992046024943, + "grad_norm": 3.3567726612091064, + "learning_rate": 0.00017070929379512723, + "loss": 0.1916, + "step": 11380 + }, + { + "epoch": 0.4397853198965211, + "grad_norm": 0.5970168709754944, + "learning_rate": 0.00017068355277552545, + "loss": 0.6096, + "step": 11390 + }, + { + "epoch": 0.4401714351905479, + "grad_norm": 2.880949020385742, + "learning_rate": 0.00017065781175592366, + "loss": 0.3566, + "step": 11400 + }, + { + "epoch": 0.4405575504845747, + "grad_norm": 0.5541375279426575, + "learning_rate": 0.00017063207073632187, + "loss": 0.4147, + "step": 11410 + }, + { + "epoch": 0.4409436657786015, + "grad_norm": 4.519477844238281, + "learning_rate": 0.00017060632971672009, + "loss": 0.4912, + "step": 11420 + }, + { + "epoch": 0.44132978107262827, + "grad_norm": 1.9959009885787964, + "learning_rate": 0.0001705805886971183, + "loss": 0.295, + "step": 11430 + }, + { + "epoch": 0.4417158963666551, + "grad_norm": 3.843033790588379, + "learning_rate": 0.0001705548476775165, + "loss": 0.2451, + "step": 11440 + }, + { + "epoch": 0.44210201166068186, + "grad_norm": 1.0480101108551025, + "learning_rate": 0.00017052910665791473, + "loss": 0.3113, + "step": 11450 + }, + { + "epoch": 0.4424881269547087, + "grad_norm": 2.0960068702697754, + "learning_rate": 0.00017050336563831294, + "loss": 0.5959, + "step": 11460 + }, + { + "epoch": 0.44287424224873545, + "grad_norm": 1.3062267303466797, + "learning_rate": 0.00017047762461871115, + "loss": 0.3812, + "step": 11470 + }, + { + "epoch": 0.4432603575427623, + "grad_norm": 2.2563138008117676, + "learning_rate": 0.00017045188359910937, + "loss": 0.303, + "step": 11480 + }, + { + "epoch": 0.44364647283678904, + "grad_norm": 1.518556833267212, + "learning_rate": 0.00017042614257950758, + "loss": 0.4349, + "step": 11490 + }, + { + "epoch": 0.44403258813081586, + "grad_norm": 5.704294681549072, + "learning_rate": 0.0001704004015599058, + "loss": 0.3162, + "step": 11500 + }, + { + "epoch": 0.44441870342484263, + "grad_norm": 3.3146274089813232, + "learning_rate": 0.000170374660540304, + "loss": 0.644, + "step": 11510 + }, + { + "epoch": 0.44480481871886945, + "grad_norm": 2.285374879837036, + "learning_rate": 0.00017034891952070222, + "loss": 0.3718, + "step": 11520 + }, + { + "epoch": 0.4451909340128963, + "grad_norm": 0.07299748063087463, + "learning_rate": 0.00017032317850110043, + "loss": 0.3093, + "step": 11530 + }, + { + "epoch": 0.44557704930692305, + "grad_norm": 4.159457683563232, + "learning_rate": 0.00017029743748149865, + "loss": 0.4074, + "step": 11540 + }, + { + "epoch": 0.44596316460094987, + "grad_norm": 2.2241604328155518, + "learning_rate": 0.00017027169646189686, + "loss": 0.2908, + "step": 11550 + }, + { + "epoch": 0.44634927989497664, + "grad_norm": 2.2086968421936035, + "learning_rate": 0.00017024595544229507, + "loss": 0.3352, + "step": 11560 + }, + { + "epoch": 0.44673539518900346, + "grad_norm": 0.2843379080295563, + "learning_rate": 0.00017022021442269329, + "loss": 0.2548, + "step": 11570 + }, + { + "epoch": 0.4471215104830302, + "grad_norm": 0.9805948734283447, + "learning_rate": 0.00017019447340309153, + "loss": 0.4134, + "step": 11580 + }, + { + "epoch": 0.44750762577705705, + "grad_norm": 2.3669955730438232, + "learning_rate": 0.0001701687323834897, + "loss": 0.3803, + "step": 11590 + }, + { + "epoch": 0.4478937410710838, + "grad_norm": 2.4062774181365967, + "learning_rate": 0.00017014299136388792, + "loss": 0.4391, + "step": 11600 + }, + { + "epoch": 0.44827985636511064, + "grad_norm": 0.4205828607082367, + "learning_rate": 0.00017011725034428614, + "loss": 0.2793, + "step": 11610 + }, + { + "epoch": 0.4486659716591374, + "grad_norm": 1.3258132934570312, + "learning_rate": 0.00017009150932468435, + "loss": 0.4219, + "step": 11620 + }, + { + "epoch": 0.44905208695316423, + "grad_norm": 3.250332832336426, + "learning_rate": 0.00017006576830508256, + "loss": 0.3638, + "step": 11630 + }, + { + "epoch": 0.449438202247191, + "grad_norm": 1.9546891450881958, + "learning_rate": 0.00017004002728548078, + "loss": 0.5473, + "step": 11640 + }, + { + "epoch": 0.4498243175412178, + "grad_norm": 0.6501532793045044, + "learning_rate": 0.00017001428626587902, + "loss": 0.197, + "step": 11650 + }, + { + "epoch": 0.4502104328352446, + "grad_norm": 0.7169322967529297, + "learning_rate": 0.0001699885452462772, + "loss": 0.2846, + "step": 11660 + }, + { + "epoch": 0.4505965481292714, + "grad_norm": 1.4476008415222168, + "learning_rate": 0.00016996280422667545, + "loss": 0.3601, + "step": 11670 + }, + { + "epoch": 0.4509826634232982, + "grad_norm": 1.562485933303833, + "learning_rate": 0.00016993706320707363, + "loss": 0.2876, + "step": 11680 + }, + { + "epoch": 0.451368778717325, + "grad_norm": 4.490612983703613, + "learning_rate": 0.00016991132218747184, + "loss": 0.5445, + "step": 11690 + }, + { + "epoch": 0.45175489401135177, + "grad_norm": 0.9045882225036621, + "learning_rate": 0.00016988558116787009, + "loss": 0.149, + "step": 11700 + }, + { + "epoch": 0.4521410093053786, + "grad_norm": 2.2920546531677246, + "learning_rate": 0.00016985984014826827, + "loss": 0.3922, + "step": 11710 + }, + { + "epoch": 0.45252712459940536, + "grad_norm": 4.459114074707031, + "learning_rate": 0.0001698340991286665, + "loss": 0.3386, + "step": 11720 + }, + { + "epoch": 0.4529132398934322, + "grad_norm": 3.0614171028137207, + "learning_rate": 0.0001698083581090647, + "loss": 0.3618, + "step": 11730 + }, + { + "epoch": 0.45329935518745895, + "grad_norm": 1.0597162246704102, + "learning_rate": 0.00016978261708946294, + "loss": 0.2012, + "step": 11740 + }, + { + "epoch": 0.4536854704814858, + "grad_norm": 0.9142243266105652, + "learning_rate": 0.00016975687606986112, + "loss": 0.3716, + "step": 11750 + }, + { + "epoch": 0.45407158577551254, + "grad_norm": 1.7402280569076538, + "learning_rate": 0.00016973113505025934, + "loss": 0.3046, + "step": 11760 + }, + { + "epoch": 0.45445770106953937, + "grad_norm": 0.7389079928398132, + "learning_rate": 0.00016970539403065758, + "loss": 0.287, + "step": 11770 + }, + { + "epoch": 0.45484381636356613, + "grad_norm": 2.0596721172332764, + "learning_rate": 0.00016967965301105576, + "loss": 0.5124, + "step": 11780 + }, + { + "epoch": 0.45522993165759296, + "grad_norm": 0.26743578910827637, + "learning_rate": 0.000169653911991454, + "loss": 0.6725, + "step": 11790 + }, + { + "epoch": 0.4556160469516198, + "grad_norm": 2.812441825866699, + "learning_rate": 0.0001696281709718522, + "loss": 0.4545, + "step": 11800 + }, + { + "epoch": 0.45600216224564655, + "grad_norm": 2.9857542514801025, + "learning_rate": 0.00016960242995225043, + "loss": 0.2865, + "step": 11810 + }, + { + "epoch": 0.45638827753967337, + "grad_norm": 0.4560181200504303, + "learning_rate": 0.00016957668893264862, + "loss": 0.1623, + "step": 11820 + }, + { + "epoch": 0.45677439283370014, + "grad_norm": 2.2492222785949707, + "learning_rate": 0.00016955094791304683, + "loss": 0.2813, + "step": 11830 + }, + { + "epoch": 0.45716050812772696, + "grad_norm": 1.0125524997711182, + "learning_rate": 0.00016952520689344507, + "loss": 0.2224, + "step": 11840 + }, + { + "epoch": 0.45754662342175373, + "grad_norm": 0.4849410057067871, + "learning_rate": 0.00016949946587384326, + "loss": 0.4838, + "step": 11850 + }, + { + "epoch": 0.45793273871578055, + "grad_norm": 2.313490390777588, + "learning_rate": 0.0001694737248542415, + "loss": 0.3855, + "step": 11860 + }, + { + "epoch": 0.4583188540098073, + "grad_norm": 2.3526558876037598, + "learning_rate": 0.00016944798383463968, + "loss": 0.3081, + "step": 11870 + }, + { + "epoch": 0.45870496930383414, + "grad_norm": 0.5723626613616943, + "learning_rate": 0.00016942224281503792, + "loss": 0.3126, + "step": 11880 + }, + { + "epoch": 0.4590910845978609, + "grad_norm": 4.130553722381592, + "learning_rate": 0.00016939650179543614, + "loss": 0.2295, + "step": 11890 + }, + { + "epoch": 0.45947719989188773, + "grad_norm": 1.3537687063217163, + "learning_rate": 0.00016937076077583432, + "loss": 0.3488, + "step": 11900 + }, + { + "epoch": 0.4598633151859145, + "grad_norm": 1.843441128730774, + "learning_rate": 0.00016934501975623256, + "loss": 0.6298, + "step": 11910 + }, + { + "epoch": 0.4602494304799413, + "grad_norm": 4.685494422912598, + "learning_rate": 0.00016931927873663075, + "loss": 0.4929, + "step": 11920 + }, + { + "epoch": 0.4606355457739681, + "grad_norm": 1.5470925569534302, + "learning_rate": 0.000169293537717029, + "loss": 0.3214, + "step": 11930 + }, + { + "epoch": 0.4610216610679949, + "grad_norm": 2.0928761959075928, + "learning_rate": 0.00016926779669742718, + "loss": 0.317, + "step": 11940 + }, + { + "epoch": 0.4614077763620217, + "grad_norm": 1.0828526020050049, + "learning_rate": 0.00016924205567782542, + "loss": 0.4204, + "step": 11950 + }, + { + "epoch": 0.4617938916560485, + "grad_norm": 2.2898383140563965, + "learning_rate": 0.00016921631465822363, + "loss": 0.4149, + "step": 11960 + }, + { + "epoch": 0.4621800069500753, + "grad_norm": 1.1590880155563354, + "learning_rate": 0.00016919057363862182, + "loss": 0.2372, + "step": 11970 + }, + { + "epoch": 0.4625661222441021, + "grad_norm": 0.6889861226081848, + "learning_rate": 0.00016916483261902006, + "loss": 0.4905, + "step": 11980 + }, + { + "epoch": 0.46295223753812886, + "grad_norm": 0.7974410653114319, + "learning_rate": 0.00016913909159941824, + "loss": 0.3451, + "step": 11990 + }, + { + "epoch": 0.4633383528321557, + "grad_norm": 3.254754066467285, + "learning_rate": 0.00016911335057981648, + "loss": 0.4018, + "step": 12000 + }, + { + "epoch": 0.46372446812618245, + "grad_norm": 2.660983085632324, + "learning_rate": 0.0001690876095602147, + "loss": 0.2112, + "step": 12010 + }, + { + "epoch": 0.4641105834202093, + "grad_norm": 1.1055381298065186, + "learning_rate": 0.0001690618685406129, + "loss": 0.2994, + "step": 12020 + }, + { + "epoch": 0.46449669871423604, + "grad_norm": 0.26787269115448, + "learning_rate": 0.00016903612752101112, + "loss": 0.2744, + "step": 12030 + }, + { + "epoch": 0.46488281400826287, + "grad_norm": 0.8832791447639465, + "learning_rate": 0.0001690103865014093, + "loss": 0.4357, + "step": 12040 + }, + { + "epoch": 0.4652689293022897, + "grad_norm": 1.0007046461105347, + "learning_rate": 0.00016898464548180755, + "loss": 0.2873, + "step": 12050 + }, + { + "epoch": 0.46565504459631646, + "grad_norm": 2.0366733074188232, + "learning_rate": 0.00016895890446220574, + "loss": 0.3016, + "step": 12060 + }, + { + "epoch": 0.4660411598903433, + "grad_norm": 2.557812213897705, + "learning_rate": 0.00016893316344260398, + "loss": 0.6375, + "step": 12070 + }, + { + "epoch": 0.46642727518437005, + "grad_norm": 1.9911783933639526, + "learning_rate": 0.0001689074224230022, + "loss": 0.5568, + "step": 12080 + }, + { + "epoch": 0.4668133904783969, + "grad_norm": 0.6905809044837952, + "learning_rate": 0.0001688816814034004, + "loss": 0.3237, + "step": 12090 + }, + { + "epoch": 0.46719950577242364, + "grad_norm": 2.3139610290527344, + "learning_rate": 0.00016885594038379862, + "loss": 0.3656, + "step": 12100 + }, + { + "epoch": 0.46758562106645046, + "grad_norm": 1.580295443534851, + "learning_rate": 0.0001688301993641968, + "loss": 0.3134, + "step": 12110 + }, + { + "epoch": 0.46797173636047723, + "grad_norm": 1.1115995645523071, + "learning_rate": 0.00016880445834459504, + "loss": 0.6167, + "step": 12120 + }, + { + "epoch": 0.46835785165450405, + "grad_norm": 3.0928075313568115, + "learning_rate": 0.00016877871732499323, + "loss": 0.3922, + "step": 12130 + }, + { + "epoch": 0.4687439669485308, + "grad_norm": 2.277163028717041, + "learning_rate": 0.00016875297630539147, + "loss": 0.2395, + "step": 12140 + }, + { + "epoch": 0.46913008224255764, + "grad_norm": 1.6578807830810547, + "learning_rate": 0.00016872723528578968, + "loss": 0.3502, + "step": 12150 + }, + { + "epoch": 0.4695161975365844, + "grad_norm": 2.9669971466064453, + "learning_rate": 0.0001687014942661879, + "loss": 0.3278, + "step": 12160 + }, + { + "epoch": 0.46990231283061124, + "grad_norm": 3.344914674758911, + "learning_rate": 0.0001686757532465861, + "loss": 0.4856, + "step": 12170 + }, + { + "epoch": 0.470288428124638, + "grad_norm": 1.512459635734558, + "learning_rate": 0.00016865001222698432, + "loss": 0.3113, + "step": 12180 + }, + { + "epoch": 0.4706745434186648, + "grad_norm": 1.8372429609298706, + "learning_rate": 0.00016862427120738254, + "loss": 0.3534, + "step": 12190 + }, + { + "epoch": 0.4710606587126916, + "grad_norm": 2.5016984939575195, + "learning_rate": 0.00016859853018778075, + "loss": 0.2129, + "step": 12200 + }, + { + "epoch": 0.4714467740067184, + "grad_norm": 2.083526134490967, + "learning_rate": 0.00016857278916817896, + "loss": 0.3011, + "step": 12210 + }, + { + "epoch": 0.4718328893007452, + "grad_norm": 3.3518013954162598, + "learning_rate": 0.00016854704814857718, + "loss": 0.2652, + "step": 12220 + }, + { + "epoch": 0.472219004594772, + "grad_norm": 0.34668633341789246, + "learning_rate": 0.0001685213071289754, + "loss": 0.3655, + "step": 12230 + }, + { + "epoch": 0.4726051198887988, + "grad_norm": 1.0674203634262085, + "learning_rate": 0.0001684955661093736, + "loss": 0.3217, + "step": 12240 + }, + { + "epoch": 0.4729912351828256, + "grad_norm": 2.9859087467193604, + "learning_rate": 0.00016846982508977182, + "loss": 0.447, + "step": 12250 + }, + { + "epoch": 0.47337735047685237, + "grad_norm": 0.33996835350990295, + "learning_rate": 0.00016844408407017003, + "loss": 0.1915, + "step": 12260 + }, + { + "epoch": 0.4737634657708792, + "grad_norm": 0.10328155755996704, + "learning_rate": 0.00016841834305056824, + "loss": 0.2588, + "step": 12270 + }, + { + "epoch": 0.47414958106490596, + "grad_norm": 2.0321199893951416, + "learning_rate": 0.00016839260203096646, + "loss": 0.4788, + "step": 12280 + }, + { + "epoch": 0.4745356963589328, + "grad_norm": 1.5690089464187622, + "learning_rate": 0.00016836686101136467, + "loss": 0.5198, + "step": 12290 + }, + { + "epoch": 0.47492181165295955, + "grad_norm": 1.989465355873108, + "learning_rate": 0.00016834111999176288, + "loss": 0.2554, + "step": 12300 + }, + { + "epoch": 0.47530792694698637, + "grad_norm": 1.574174404144287, + "learning_rate": 0.0001683153789721611, + "loss": 0.3703, + "step": 12310 + }, + { + "epoch": 0.4756940422410132, + "grad_norm": 2.726776599884033, + "learning_rate": 0.0001682896379525593, + "loss": 0.7426, + "step": 12320 + }, + { + "epoch": 0.47608015753503996, + "grad_norm": 0.3101334571838379, + "learning_rate": 0.00016826389693295752, + "loss": 0.1861, + "step": 12330 + }, + { + "epoch": 0.4764662728290668, + "grad_norm": 5.30327033996582, + "learning_rate": 0.00016823815591335574, + "loss": 0.4261, + "step": 12340 + }, + { + "epoch": 0.47685238812309355, + "grad_norm": 0.5553661584854126, + "learning_rate": 0.00016821241489375395, + "loss": 0.7326, + "step": 12350 + }, + { + "epoch": 0.4772385034171204, + "grad_norm": 2.2244138717651367, + "learning_rate": 0.00016818667387415216, + "loss": 0.5406, + "step": 12360 + }, + { + "epoch": 0.47762461871114714, + "grad_norm": 0.5314281582832336, + "learning_rate": 0.00016816093285455038, + "loss": 0.3765, + "step": 12370 + }, + { + "epoch": 0.47801073400517397, + "grad_norm": 3.644477128982544, + "learning_rate": 0.0001681351918349486, + "loss": 0.5519, + "step": 12380 + }, + { + "epoch": 0.47839684929920073, + "grad_norm": 1.0556128025054932, + "learning_rate": 0.0001681094508153468, + "loss": 0.2128, + "step": 12390 + }, + { + "epoch": 0.47878296459322756, + "grad_norm": 2.7807135581970215, + "learning_rate": 0.00016808370979574502, + "loss": 0.1564, + "step": 12400 + }, + { + "epoch": 0.4791690798872543, + "grad_norm": 0.7862847447395325, + "learning_rate": 0.00016805796877614323, + "loss": 0.35, + "step": 12410 + }, + { + "epoch": 0.47955519518128115, + "grad_norm": 2.7445156574249268, + "learning_rate": 0.00016803222775654144, + "loss": 0.4167, + "step": 12420 + }, + { + "epoch": 0.4799413104753079, + "grad_norm": 0.6173526644706726, + "learning_rate": 0.00016800648673693966, + "loss": 0.3669, + "step": 12430 + }, + { + "epoch": 0.48032742576933474, + "grad_norm": 0.3762228488922119, + "learning_rate": 0.00016798074571733787, + "loss": 0.179, + "step": 12440 + }, + { + "epoch": 0.4807135410633615, + "grad_norm": 1.2806499004364014, + "learning_rate": 0.00016795500469773608, + "loss": 0.3926, + "step": 12450 + }, + { + "epoch": 0.48109965635738833, + "grad_norm": 1.0579105615615845, + "learning_rate": 0.0001679292636781343, + "loss": 0.3329, + "step": 12460 + }, + { + "epoch": 0.4814857716514151, + "grad_norm": 1.6502699851989746, + "learning_rate": 0.0001679035226585325, + "loss": 0.3149, + "step": 12470 + }, + { + "epoch": 0.4818718869454419, + "grad_norm": 0.8876189589500427, + "learning_rate": 0.00016787778163893072, + "loss": 0.3262, + "step": 12480 + }, + { + "epoch": 0.4822580022394687, + "grad_norm": 2.4398770332336426, + "learning_rate": 0.00016785204061932894, + "loss": 0.3834, + "step": 12490 + }, + { + "epoch": 0.4826441175334955, + "grad_norm": 1.275148868560791, + "learning_rate": 0.00016782629959972715, + "loss": 0.3749, + "step": 12500 + }, + { + "epoch": 0.4830302328275223, + "grad_norm": 2.1397783756256104, + "learning_rate": 0.0001678005585801254, + "loss": 0.4726, + "step": 12510 + }, + { + "epoch": 0.4834163481215491, + "grad_norm": 1.6431002616882324, + "learning_rate": 0.00016777481756052358, + "loss": 0.3537, + "step": 12520 + }, + { + "epoch": 0.48380246341557587, + "grad_norm": 1.1193108558654785, + "learning_rate": 0.0001677490765409218, + "loss": 0.3182, + "step": 12530 + }, + { + "epoch": 0.4841885787096027, + "grad_norm": 1.365897297859192, + "learning_rate": 0.00016772333552132, + "loss": 0.3455, + "step": 12540 + }, + { + "epoch": 0.48457469400362946, + "grad_norm": 0.9035172462463379, + "learning_rate": 0.00016769759450171822, + "loss": 0.2268, + "step": 12550 + }, + { + "epoch": 0.4849608092976563, + "grad_norm": 0.15492293238639832, + "learning_rate": 0.00016767185348211643, + "loss": 0.319, + "step": 12560 + }, + { + "epoch": 0.4853469245916831, + "grad_norm": 0.19394727051258087, + "learning_rate": 0.00016764611246251464, + "loss": 0.2073, + "step": 12570 + }, + { + "epoch": 0.48573303988570987, + "grad_norm": 1.4406816959381104, + "learning_rate": 0.00016762037144291288, + "loss": 0.4538, + "step": 12580 + }, + { + "epoch": 0.4861191551797367, + "grad_norm": 2.741548538208008, + "learning_rate": 0.00016759463042331107, + "loss": 0.536, + "step": 12590 + }, + { + "epoch": 0.48650527047376346, + "grad_norm": 0.5203917622566223, + "learning_rate": 0.00016756888940370928, + "loss": 0.2484, + "step": 12600 + }, + { + "epoch": 0.4868913857677903, + "grad_norm": 0.9666195511817932, + "learning_rate": 0.0001675431483841075, + "loss": 0.3834, + "step": 12610 + }, + { + "epoch": 0.48727750106181705, + "grad_norm": 1.7210304737091064, + "learning_rate": 0.0001675174073645057, + "loss": 0.3223, + "step": 12620 + }, + { + "epoch": 0.4876636163558439, + "grad_norm": 0.6125622987747192, + "learning_rate": 0.00016749166634490392, + "loss": 0.394, + "step": 12630 + }, + { + "epoch": 0.48804973164987064, + "grad_norm": 1.9612951278686523, + "learning_rate": 0.00016746592532530214, + "loss": 0.3367, + "step": 12640 + }, + { + "epoch": 0.48843584694389747, + "grad_norm": 1.5395468473434448, + "learning_rate": 0.00016744018430570038, + "loss": 0.5441, + "step": 12650 + }, + { + "epoch": 0.48882196223792423, + "grad_norm": 0.8706358671188354, + "learning_rate": 0.00016741444328609856, + "loss": 0.422, + "step": 12660 + }, + { + "epoch": 0.48920807753195106, + "grad_norm": 4.338102340698242, + "learning_rate": 0.00016738870226649678, + "loss": 0.3932, + "step": 12670 + }, + { + "epoch": 0.4895941928259778, + "grad_norm": 0.9501354694366455, + "learning_rate": 0.000167362961246895, + "loss": 0.563, + "step": 12680 + }, + { + "epoch": 0.48998030812000465, + "grad_norm": 1.3146884441375732, + "learning_rate": 0.0001673372202272932, + "loss": 0.5116, + "step": 12690 + }, + { + "epoch": 0.4903664234140314, + "grad_norm": 2.144622564315796, + "learning_rate": 0.00016731147920769144, + "loss": 0.4792, + "step": 12700 + }, + { + "epoch": 0.49075253870805824, + "grad_norm": 0.9871418476104736, + "learning_rate": 0.00016728573818808963, + "loss": 0.3802, + "step": 12710 + }, + { + "epoch": 0.491138654002085, + "grad_norm": 1.710766315460205, + "learning_rate": 0.00016725999716848787, + "loss": 0.2859, + "step": 12720 + }, + { + "epoch": 0.49152476929611183, + "grad_norm": 2.528146505355835, + "learning_rate": 0.00016723425614888606, + "loss": 0.4565, + "step": 12730 + }, + { + "epoch": 0.4919108845901386, + "grad_norm": 0.8471786379814148, + "learning_rate": 0.00016720851512928427, + "loss": 0.3455, + "step": 12740 + }, + { + "epoch": 0.4922969998841654, + "grad_norm": 2.6623692512512207, + "learning_rate": 0.00016718277410968248, + "loss": 0.3687, + "step": 12750 + }, + { + "epoch": 0.4926831151781922, + "grad_norm": 2.9324758052825928, + "learning_rate": 0.0001671570330900807, + "loss": 0.2652, + "step": 12760 + }, + { + "epoch": 0.493069230472219, + "grad_norm": 1.7955294847488403, + "learning_rate": 0.00016713129207047894, + "loss": 0.3156, + "step": 12770 + }, + { + "epoch": 0.4934553457662458, + "grad_norm": 0.9923033118247986, + "learning_rate": 0.00016710555105087712, + "loss": 0.3042, + "step": 12780 + }, + { + "epoch": 0.4938414610602726, + "grad_norm": 0.9309022426605225, + "learning_rate": 0.00016707981003127536, + "loss": 0.2742, + "step": 12790 + }, + { + "epoch": 0.49422757635429937, + "grad_norm": 4.864802360534668, + "learning_rate": 0.00016705406901167355, + "loss": 0.6145, + "step": 12800 + }, + { + "epoch": 0.4946136916483262, + "grad_norm": 2.4508230686187744, + "learning_rate": 0.00016702832799207176, + "loss": 0.4218, + "step": 12810 + }, + { + "epoch": 0.49499980694235296, + "grad_norm": 2.0317444801330566, + "learning_rate": 0.00016700258697246998, + "loss": 0.5682, + "step": 12820 + }, + { + "epoch": 0.4953859222363798, + "grad_norm": 4.89669942855835, + "learning_rate": 0.0001669768459528682, + "loss": 0.5654, + "step": 12830 + }, + { + "epoch": 0.4957720375304066, + "grad_norm": 0.602165162563324, + "learning_rate": 0.00016695110493326643, + "loss": 0.2719, + "step": 12840 + }, + { + "epoch": 0.4961581528244334, + "grad_norm": 1.1574476957321167, + "learning_rate": 0.00016692536391366462, + "loss": 0.3872, + "step": 12850 + }, + { + "epoch": 0.4965442681184602, + "grad_norm": 0.4792019724845886, + "learning_rate": 0.00016689962289406286, + "loss": 0.2671, + "step": 12860 + }, + { + "epoch": 0.49693038341248696, + "grad_norm": 1.4611676931381226, + "learning_rate": 0.00016687388187446104, + "loss": 0.3378, + "step": 12870 + }, + { + "epoch": 0.4973164987065138, + "grad_norm": 1.183975338935852, + "learning_rate": 0.00016684814085485928, + "loss": 0.2645, + "step": 12880 + }, + { + "epoch": 0.49770261400054056, + "grad_norm": 2.1447482109069824, + "learning_rate": 0.0001668223998352575, + "loss": 0.1678, + "step": 12890 + }, + { + "epoch": 0.4980887292945674, + "grad_norm": 1.5829964876174927, + "learning_rate": 0.00016679665881565568, + "loss": 0.4113, + "step": 12900 + }, + { + "epoch": 0.49847484458859415, + "grad_norm": 1.329871654510498, + "learning_rate": 0.00016677091779605392, + "loss": 0.2639, + "step": 12910 + }, + { + "epoch": 0.49886095988262097, + "grad_norm": 4.797327518463135, + "learning_rate": 0.0001667451767764521, + "loss": 0.3396, + "step": 12920 + }, + { + "epoch": 0.49924707517664774, + "grad_norm": 2.5864250659942627, + "learning_rate": 0.00016671943575685035, + "loss": 0.4143, + "step": 12930 + }, + { + "epoch": 0.49963319047067456, + "grad_norm": 0.05073557794094086, + "learning_rate": 0.00016669369473724854, + "loss": 0.231, + "step": 12940 + }, + { + "epoch": 0.5000193057647013, + "grad_norm": 0.9530317783355713, + "learning_rate": 0.00016666795371764678, + "loss": 0.3702, + "step": 12950 + }, + { + "epoch": 0.5004054210587281, + "grad_norm": 1.3573989868164062, + "learning_rate": 0.000166642212698045, + "loss": 0.2095, + "step": 12960 + }, + { + "epoch": 0.500791536352755, + "grad_norm": 3.2758514881134033, + "learning_rate": 0.00016661647167844318, + "loss": 0.3901, + "step": 12970 + }, + { + "epoch": 0.5011776516467817, + "grad_norm": 2.359602212905884, + "learning_rate": 0.00016659073065884142, + "loss": 0.6232, + "step": 12980 + }, + { + "epoch": 0.5015637669408085, + "grad_norm": 0.5743809938430786, + "learning_rate": 0.0001665649896392396, + "loss": 0.431, + "step": 12990 + }, + { + "epoch": 0.5019498822348353, + "grad_norm": 2.229215621948242, + "learning_rate": 0.00016653924861963784, + "loss": 0.33, + "step": 13000 + }, + { + "epoch": 0.5023359975288622, + "grad_norm": 3.1233408451080322, + "learning_rate": 0.00016651350760003606, + "loss": 0.3527, + "step": 13010 + }, + { + "epoch": 0.5027221128228889, + "grad_norm": 0.4846508204936981, + "learning_rate": 0.00016648776658043427, + "loss": 0.4514, + "step": 13020 + }, + { + "epoch": 0.5031082281169157, + "grad_norm": 0.807447612285614, + "learning_rate": 0.00016646202556083248, + "loss": 0.3643, + "step": 13030 + }, + { + "epoch": 0.5034943434109425, + "grad_norm": 3.226194143295288, + "learning_rate": 0.00016643628454123067, + "loss": 0.4803, + "step": 13040 + }, + { + "epoch": 0.5038804587049693, + "grad_norm": 6.127805233001709, + "learning_rate": 0.0001664105435216289, + "loss": 0.3501, + "step": 13050 + }, + { + "epoch": 0.5042665739989961, + "grad_norm": 1.3367782831192017, + "learning_rate": 0.0001663848025020271, + "loss": 0.2259, + "step": 13060 + }, + { + "epoch": 0.5046526892930229, + "grad_norm": 1.112602710723877, + "learning_rate": 0.00016635906148242534, + "loss": 0.3868, + "step": 13070 + }, + { + "epoch": 0.5050388045870496, + "grad_norm": 3.188282012939453, + "learning_rate": 0.00016633332046282355, + "loss": 0.2783, + "step": 13080 + }, + { + "epoch": 0.5054249198810765, + "grad_norm": 1.5972063541412354, + "learning_rate": 0.00016630757944322176, + "loss": 0.4661, + "step": 13090 + }, + { + "epoch": 0.5058110351751033, + "grad_norm": 0.8210055232048035, + "learning_rate": 0.00016628183842361998, + "loss": 0.2536, + "step": 13100 + }, + { + "epoch": 0.50619715046913, + "grad_norm": 2.279244899749756, + "learning_rate": 0.00016625609740401816, + "loss": 0.3158, + "step": 13110 + }, + { + "epoch": 0.5065832657631569, + "grad_norm": 2.9017488956451416, + "learning_rate": 0.0001662303563844164, + "loss": 0.4209, + "step": 13120 + }, + { + "epoch": 0.5069693810571837, + "grad_norm": 0.9567920565605164, + "learning_rate": 0.0001662046153648146, + "loss": 0.3869, + "step": 13130 + }, + { + "epoch": 0.5073554963512105, + "grad_norm": 1.3605408668518066, + "learning_rate": 0.00016617887434521283, + "loss": 0.3529, + "step": 13140 + }, + { + "epoch": 0.5077416116452372, + "grad_norm": 0.502921998500824, + "learning_rate": 0.00016615313332561104, + "loss": 0.2584, + "step": 13150 + }, + { + "epoch": 0.5081277269392641, + "grad_norm": 2.9092366695404053, + "learning_rate": 0.00016612739230600926, + "loss": 0.3494, + "step": 13160 + }, + { + "epoch": 0.5085138422332909, + "grad_norm": 0.28300145268440247, + "learning_rate": 0.00016610165128640747, + "loss": 0.3514, + "step": 13170 + }, + { + "epoch": 0.5088999575273176, + "grad_norm": 1.326134204864502, + "learning_rate": 0.00016607591026680566, + "loss": 0.1593, + "step": 13180 + }, + { + "epoch": 0.5092860728213444, + "grad_norm": 2.1484436988830566, + "learning_rate": 0.0001660501692472039, + "loss": 0.4468, + "step": 13190 + }, + { + "epoch": 0.5096721881153713, + "grad_norm": 1.9255646467208862, + "learning_rate": 0.0001660244282276021, + "loss": 0.3937, + "step": 13200 + }, + { + "epoch": 0.5100583034093981, + "grad_norm": 0.01497764140367508, + "learning_rate": 0.00016599868720800032, + "loss": 0.2795, + "step": 13210 + }, + { + "epoch": 0.5104444187034248, + "grad_norm": 0.992023766040802, + "learning_rate": 0.00016597294618839854, + "loss": 0.2389, + "step": 13220 + }, + { + "epoch": 0.5108305339974516, + "grad_norm": 1.517337441444397, + "learning_rate": 0.00016594720516879675, + "loss": 0.2821, + "step": 13230 + }, + { + "epoch": 0.5112166492914785, + "grad_norm": 2.355637311935425, + "learning_rate": 0.00016592146414919496, + "loss": 0.5727, + "step": 13240 + }, + { + "epoch": 0.5116027645855052, + "grad_norm": 2.450536012649536, + "learning_rate": 0.00016589572312959315, + "loss": 0.4887, + "step": 13250 + }, + { + "epoch": 0.511988879879532, + "grad_norm": 1.2966598272323608, + "learning_rate": 0.0001658699821099914, + "loss": 0.3605, + "step": 13260 + }, + { + "epoch": 0.5123749951735588, + "grad_norm": 1.3981765508651733, + "learning_rate": 0.0001658442410903896, + "loss": 0.5329, + "step": 13270 + }, + { + "epoch": 0.5127611104675857, + "grad_norm": 0.7260739803314209, + "learning_rate": 0.00016581850007078782, + "loss": 0.3447, + "step": 13280 + }, + { + "epoch": 0.5131472257616124, + "grad_norm": 1.1348093748092651, + "learning_rate": 0.00016579275905118603, + "loss": 0.3086, + "step": 13290 + }, + { + "epoch": 0.5135333410556392, + "grad_norm": 1.3015291690826416, + "learning_rate": 0.00016576701803158424, + "loss": 0.3411, + "step": 13300 + }, + { + "epoch": 0.513919456349666, + "grad_norm": 2.352766990661621, + "learning_rate": 0.00016574127701198246, + "loss": 0.1617, + "step": 13310 + }, + { + "epoch": 0.5143055716436928, + "grad_norm": 2.0096113681793213, + "learning_rate": 0.00016571553599238067, + "loss": 0.3168, + "step": 13320 + }, + { + "epoch": 0.5146916869377196, + "grad_norm": 0.8163488507270813, + "learning_rate": 0.00016568979497277888, + "loss": 0.3318, + "step": 13330 + }, + { + "epoch": 0.5150778022317464, + "grad_norm": 0.8218249082565308, + "learning_rate": 0.0001656640539531771, + "loss": 0.2303, + "step": 13340 + }, + { + "epoch": 0.5154639175257731, + "grad_norm": 3.5414462089538574, + "learning_rate": 0.0001656383129335753, + "loss": 0.2295, + "step": 13350 + }, + { + "epoch": 0.5158500328198, + "grad_norm": 2.176178216934204, + "learning_rate": 0.00016561257191397352, + "loss": 0.5767, + "step": 13360 + }, + { + "epoch": 0.5162361481138268, + "grad_norm": 1.4649319648742676, + "learning_rate": 0.00016558683089437174, + "loss": 0.174, + "step": 13370 + }, + { + "epoch": 0.5166222634078536, + "grad_norm": 2.107895851135254, + "learning_rate": 0.00016556108987476995, + "loss": 0.3444, + "step": 13380 + }, + { + "epoch": 0.5170083787018804, + "grad_norm": 1.3116638660430908, + "learning_rate": 0.00016553534885516816, + "loss": 0.2462, + "step": 13390 + }, + { + "epoch": 0.5173944939959072, + "grad_norm": 2.867553949356079, + "learning_rate": 0.00016550960783556638, + "loss": 0.3106, + "step": 13400 + }, + { + "epoch": 0.517780609289934, + "grad_norm": 3.4331533908843994, + "learning_rate": 0.0001654838668159646, + "loss": 0.5359, + "step": 13410 + }, + { + "epoch": 0.5181667245839607, + "grad_norm": 1.8306528329849243, + "learning_rate": 0.0001654581257963628, + "loss": 0.4225, + "step": 13420 + }, + { + "epoch": 0.5185528398779876, + "grad_norm": 1.242026448249817, + "learning_rate": 0.00016543238477676102, + "loss": 0.2679, + "step": 13430 + }, + { + "epoch": 0.5189389551720144, + "grad_norm": 0.5788571834564209, + "learning_rate": 0.00016540664375715923, + "loss": 0.369, + "step": 13440 + }, + { + "epoch": 0.5193250704660411, + "grad_norm": 0.22553350031375885, + "learning_rate": 0.00016538090273755744, + "loss": 0.422, + "step": 13450 + }, + { + "epoch": 0.5197111857600679, + "grad_norm": 1.6932384967803955, + "learning_rate": 0.00016535516171795566, + "loss": 0.3127, + "step": 13460 + }, + { + "epoch": 0.5200973010540948, + "grad_norm": 1.3747683763504028, + "learning_rate": 0.00016532942069835387, + "loss": 0.2751, + "step": 13470 + }, + { + "epoch": 0.5204834163481216, + "grad_norm": 2.0508596897125244, + "learning_rate": 0.00016530367967875208, + "loss": 0.4951, + "step": 13480 + }, + { + "epoch": 0.5208695316421483, + "grad_norm": 1.044503092765808, + "learning_rate": 0.0001652779386591503, + "loss": 0.2329, + "step": 13490 + }, + { + "epoch": 0.5212556469361751, + "grad_norm": 1.6400004625320435, + "learning_rate": 0.0001652521976395485, + "loss": 0.1846, + "step": 13500 + }, + { + "epoch": 0.521641762230202, + "grad_norm": 2.234170913696289, + "learning_rate": 0.00016522645661994672, + "loss": 0.5005, + "step": 13510 + }, + { + "epoch": 0.5220278775242287, + "grad_norm": 1.8167870044708252, + "learning_rate": 0.00016520071560034493, + "loss": 0.4629, + "step": 13520 + }, + { + "epoch": 0.5224139928182555, + "grad_norm": 1.7200794219970703, + "learning_rate": 0.00016517497458074315, + "loss": 0.4546, + "step": 13530 + }, + { + "epoch": 0.5228001081122823, + "grad_norm": 2.845244884490967, + "learning_rate": 0.00016514923356114136, + "loss": 0.2211, + "step": 13540 + }, + { + "epoch": 0.5231862234063092, + "grad_norm": 5.3078389167785645, + "learning_rate": 0.00016512349254153957, + "loss": 0.4903, + "step": 13550 + }, + { + "epoch": 0.5235723387003359, + "grad_norm": 8.109561920166016, + "learning_rate": 0.0001650977515219378, + "loss": 0.3624, + "step": 13560 + }, + { + "epoch": 0.5239584539943627, + "grad_norm": 0.537749171257019, + "learning_rate": 0.000165072010502336, + "loss": 0.2262, + "step": 13570 + }, + { + "epoch": 0.5243445692883895, + "grad_norm": 2.8182802200317383, + "learning_rate": 0.00016504626948273421, + "loss": 0.4173, + "step": 13580 + }, + { + "epoch": 0.5247306845824163, + "grad_norm": 0.6623479127883911, + "learning_rate": 0.00016502052846313243, + "loss": 0.2955, + "step": 13590 + }, + { + "epoch": 0.5251167998764431, + "grad_norm": 3.027707099914551, + "learning_rate": 0.00016499478744353064, + "loss": 0.5272, + "step": 13600 + }, + { + "epoch": 0.5255029151704699, + "grad_norm": 1.6830018758773804, + "learning_rate": 0.00016496904642392885, + "loss": 0.2169, + "step": 13610 + }, + { + "epoch": 0.5258890304644968, + "grad_norm": 3.0182113647460938, + "learning_rate": 0.00016494330540432707, + "loss": 0.5747, + "step": 13620 + }, + { + "epoch": 0.5262751457585235, + "grad_norm": 1.5487585067749023, + "learning_rate": 0.00016491756438472528, + "loss": 0.2585, + "step": 13630 + }, + { + "epoch": 0.5266612610525503, + "grad_norm": 1.2525122165679932, + "learning_rate": 0.0001648918233651235, + "loss": 0.3622, + "step": 13640 + }, + { + "epoch": 0.5270473763465771, + "grad_norm": 1.9910658597946167, + "learning_rate": 0.00016486608234552174, + "loss": 0.2049, + "step": 13650 + }, + { + "epoch": 0.5274334916406039, + "grad_norm": 0.7254251837730408, + "learning_rate": 0.00016484034132591992, + "loss": 0.3101, + "step": 13660 + }, + { + "epoch": 0.5278196069346307, + "grad_norm": 0.9839001893997192, + "learning_rate": 0.00016481460030631813, + "loss": 0.4723, + "step": 13670 + }, + { + "epoch": 0.5282057222286575, + "grad_norm": 2.176529884338379, + "learning_rate": 0.00016478885928671635, + "loss": 0.3669, + "step": 13680 + }, + { + "epoch": 0.5285918375226842, + "grad_norm": 2.534996509552002, + "learning_rate": 0.00016476311826711456, + "loss": 0.551, + "step": 13690 + }, + { + "epoch": 0.5289779528167111, + "grad_norm": 2.0100669860839844, + "learning_rate": 0.0001647373772475128, + "loss": 0.445, + "step": 13700 + }, + { + "epoch": 0.5293640681107379, + "grad_norm": 0.7759265899658203, + "learning_rate": 0.000164711636227911, + "loss": 0.2839, + "step": 13710 + }, + { + "epoch": 0.5297501834047647, + "grad_norm": 3.398287057876587, + "learning_rate": 0.00016468589520830923, + "loss": 0.3497, + "step": 13720 + }, + { + "epoch": 0.5301362986987914, + "grad_norm": 2.6792221069335938, + "learning_rate": 0.00016466015418870741, + "loss": 0.3435, + "step": 13730 + }, + { + "epoch": 0.5305224139928183, + "grad_norm": 0.7382081747055054, + "learning_rate": 0.00016463441316910563, + "loss": 0.367, + "step": 13740 + }, + { + "epoch": 0.5309085292868451, + "grad_norm": 0.9496407508850098, + "learning_rate": 0.00016460867214950384, + "loss": 0.305, + "step": 13750 + }, + { + "epoch": 0.5312946445808718, + "grad_norm": 1.2950342893600464, + "learning_rate": 0.00016458293112990205, + "loss": 0.2769, + "step": 13760 + }, + { + "epoch": 0.5316807598748986, + "grad_norm": 1.1744359731674194, + "learning_rate": 0.0001645571901103003, + "loss": 0.4776, + "step": 13770 + }, + { + "epoch": 0.5320668751689255, + "grad_norm": 1.1507617235183716, + "learning_rate": 0.00016453144909069848, + "loss": 0.4486, + "step": 13780 + }, + { + "epoch": 0.5324529904629522, + "grad_norm": 3.200432300567627, + "learning_rate": 0.00016450570807109672, + "loss": 0.4144, + "step": 13790 + }, + { + "epoch": 0.532839105756979, + "grad_norm": 0.991581916809082, + "learning_rate": 0.0001644799670514949, + "loss": 0.2314, + "step": 13800 + }, + { + "epoch": 0.5332252210510058, + "grad_norm": 4.800248622894287, + "learning_rate": 0.00016445422603189312, + "loss": 0.4601, + "step": 13810 + }, + { + "epoch": 0.5336113363450327, + "grad_norm": 1.2141329050064087, + "learning_rate": 0.00016442848501229136, + "loss": 0.257, + "step": 13820 + }, + { + "epoch": 0.5339974516390594, + "grad_norm": 0.8803738951683044, + "learning_rate": 0.00016440274399268955, + "loss": 0.4645, + "step": 13830 + }, + { + "epoch": 0.5343835669330862, + "grad_norm": 1.2020646333694458, + "learning_rate": 0.0001643770029730878, + "loss": 0.3751, + "step": 13840 + }, + { + "epoch": 0.534769682227113, + "grad_norm": 0.9887505173683167, + "learning_rate": 0.00016435126195348597, + "loss": 0.2266, + "step": 13850 + }, + { + "epoch": 0.5351557975211398, + "grad_norm": 0.38067731261253357, + "learning_rate": 0.00016432552093388421, + "loss": 0.3482, + "step": 13860 + }, + { + "epoch": 0.5355419128151666, + "grad_norm": 5.429462909698486, + "learning_rate": 0.0001642997799142824, + "loss": 0.4055, + "step": 13870 + }, + { + "epoch": 0.5359280281091934, + "grad_norm": 2.197861909866333, + "learning_rate": 0.00016427403889468061, + "loss": 0.1701, + "step": 13880 + }, + { + "epoch": 0.5363141434032203, + "grad_norm": 1.467132568359375, + "learning_rate": 0.00016424829787507885, + "loss": 0.3536, + "step": 13890 + }, + { + "epoch": 0.536700258697247, + "grad_norm": 2.0702550411224365, + "learning_rate": 0.00016422255685547704, + "loss": 0.5644, + "step": 13900 + }, + { + "epoch": 0.5370863739912738, + "grad_norm": 1.4855132102966309, + "learning_rate": 0.00016419681583587528, + "loss": 0.4068, + "step": 13910 + }, + { + "epoch": 0.5374724892853006, + "grad_norm": 0.6586676239967346, + "learning_rate": 0.00016417107481627347, + "loss": 0.3709, + "step": 13920 + }, + { + "epoch": 0.5378586045793274, + "grad_norm": 0.026774466037750244, + "learning_rate": 0.0001641453337966717, + "loss": 0.2635, + "step": 13930 + }, + { + "epoch": 0.5382447198733542, + "grad_norm": 0.9848103523254395, + "learning_rate": 0.0001641195927770699, + "loss": 0.249, + "step": 13940 + }, + { + "epoch": 0.538630835167381, + "grad_norm": 3.77512526512146, + "learning_rate": 0.0001640938517574681, + "loss": 0.5356, + "step": 13950 + }, + { + "epoch": 0.5390169504614077, + "grad_norm": 5.543573379516602, + "learning_rate": 0.00016406811073786635, + "loss": 0.6471, + "step": 13960 + }, + { + "epoch": 0.5394030657554346, + "grad_norm": 0.8161652684211731, + "learning_rate": 0.00016404236971826453, + "loss": 0.5691, + "step": 13970 + }, + { + "epoch": 0.5397891810494614, + "grad_norm": 1.5539859533309937, + "learning_rate": 0.00016401662869866277, + "loss": 0.4406, + "step": 13980 + }, + { + "epoch": 0.5401752963434882, + "grad_norm": 1.3035658597946167, + "learning_rate": 0.00016399088767906096, + "loss": 0.3086, + "step": 13990 + }, + { + "epoch": 0.5405614116375149, + "grad_norm": 0.9168418645858765, + "learning_rate": 0.0001639651466594592, + "loss": 0.1581, + "step": 14000 + }, + { + "epoch": 0.5409475269315418, + "grad_norm": 1.0382287502288818, + "learning_rate": 0.00016393940563985741, + "loss": 0.4723, + "step": 14010 + }, + { + "epoch": 0.5413336422255686, + "grad_norm": 2.896981716156006, + "learning_rate": 0.0001639136646202556, + "loss": 0.2999, + "step": 14020 + }, + { + "epoch": 0.5417197575195953, + "grad_norm": 0.7354179620742798, + "learning_rate": 0.00016388792360065384, + "loss": 0.4853, + "step": 14030 + }, + { + "epoch": 0.5421058728136221, + "grad_norm": 3.221067190170288, + "learning_rate": 0.00016386218258105203, + "loss": 0.3622, + "step": 14040 + }, + { + "epoch": 0.542491988107649, + "grad_norm": 6.591146469116211, + "learning_rate": 0.00016383644156145027, + "loss": 0.5803, + "step": 14050 + }, + { + "epoch": 0.5428781034016757, + "grad_norm": 3.1521377563476562, + "learning_rate": 0.00016381070054184845, + "loss": 0.3267, + "step": 14060 + }, + { + "epoch": 0.5432642186957025, + "grad_norm": 1.7890762090682983, + "learning_rate": 0.0001637849595222467, + "loss": 0.4584, + "step": 14070 + }, + { + "epoch": 0.5436503339897293, + "grad_norm": 1.6599558591842651, + "learning_rate": 0.0001637592185026449, + "loss": 0.298, + "step": 14080 + }, + { + "epoch": 0.5440364492837562, + "grad_norm": 3.521927833557129, + "learning_rate": 0.00016373347748304312, + "loss": 0.3743, + "step": 14090 + }, + { + "epoch": 0.5444225645777829, + "grad_norm": 3.8942599296569824, + "learning_rate": 0.00016370773646344133, + "loss": 0.3254, + "step": 14100 + }, + { + "epoch": 0.5448086798718097, + "grad_norm": 2.8547496795654297, + "learning_rate": 0.00016368199544383952, + "loss": 0.4073, + "step": 14110 + }, + { + "epoch": 0.5451947951658365, + "grad_norm": 1.0060430765151978, + "learning_rate": 0.00016365625442423776, + "loss": 0.1631, + "step": 14120 + }, + { + "epoch": 0.5455809104598633, + "grad_norm": 2.2001001834869385, + "learning_rate": 0.00016363051340463595, + "loss": 0.2854, + "step": 14130 + }, + { + "epoch": 0.5459670257538901, + "grad_norm": 1.3699944019317627, + "learning_rate": 0.0001636047723850342, + "loss": 0.456, + "step": 14140 + }, + { + "epoch": 0.5463531410479169, + "grad_norm": 2.1481733322143555, + "learning_rate": 0.0001635790313654324, + "loss": 0.4085, + "step": 14150 + }, + { + "epoch": 0.5467392563419438, + "grad_norm": 0.40439683198928833, + "learning_rate": 0.00016355329034583061, + "loss": 0.3932, + "step": 14160 + }, + { + "epoch": 0.5471253716359705, + "grad_norm": 2.2773404121398926, + "learning_rate": 0.00016352754932622883, + "loss": 0.2926, + "step": 14170 + }, + { + "epoch": 0.5475114869299973, + "grad_norm": 2.2974839210510254, + "learning_rate": 0.000163501808306627, + "loss": 0.2419, + "step": 14180 + }, + { + "epoch": 0.5478976022240241, + "grad_norm": 1.0429989099502563, + "learning_rate": 0.00016347606728702525, + "loss": 0.2559, + "step": 14190 + }, + { + "epoch": 0.5482837175180509, + "grad_norm": 0.8988879919052124, + "learning_rate": 0.00016345032626742347, + "loss": 0.1677, + "step": 14200 + }, + { + "epoch": 0.5486698328120777, + "grad_norm": 1.2740018367767334, + "learning_rate": 0.00016342458524782168, + "loss": 0.2452, + "step": 14210 + }, + { + "epoch": 0.5490559481061045, + "grad_norm": 1.789467692375183, + "learning_rate": 0.0001633988442282199, + "loss": 0.2272, + "step": 14220 + }, + { + "epoch": 0.5494420634001312, + "grad_norm": 3.2136781215667725, + "learning_rate": 0.0001633731032086181, + "loss": 0.4026, + "step": 14230 + }, + { + "epoch": 0.5498281786941581, + "grad_norm": 2.4747092723846436, + "learning_rate": 0.00016334736218901632, + "loss": 0.2371, + "step": 14240 + }, + { + "epoch": 0.5502142939881849, + "grad_norm": 1.5639567375183105, + "learning_rate": 0.0001633216211694145, + "loss": 0.2801, + "step": 14250 + }, + { + "epoch": 0.5506004092822117, + "grad_norm": 3.9598312377929688, + "learning_rate": 0.00016329588014981275, + "loss": 0.2583, + "step": 14260 + }, + { + "epoch": 0.5509865245762384, + "grad_norm": 1.5873563289642334, + "learning_rate": 0.00016327013913021096, + "loss": 0.2729, + "step": 14270 + }, + { + "epoch": 0.5513726398702653, + "grad_norm": 2.2313668727874756, + "learning_rate": 0.00016324439811060917, + "loss": 0.191, + "step": 14280 + }, + { + "epoch": 0.5517587551642921, + "grad_norm": 1.6087117195129395, + "learning_rate": 0.0001632186570910074, + "loss": 0.2698, + "step": 14290 + }, + { + "epoch": 0.5521448704583188, + "grad_norm": 6.5459675788879395, + "learning_rate": 0.0001631929160714056, + "loss": 0.3632, + "step": 14300 + }, + { + "epoch": 0.5525309857523456, + "grad_norm": 1.2121779918670654, + "learning_rate": 0.00016316717505180381, + "loss": 0.4541, + "step": 14310 + }, + { + "epoch": 0.5529171010463725, + "grad_norm": 2.7277257442474365, + "learning_rate": 0.00016314143403220203, + "loss": 0.1489, + "step": 14320 + }, + { + "epoch": 0.5533032163403993, + "grad_norm": 2.2566685676574707, + "learning_rate": 0.00016311569301260024, + "loss": 0.1838, + "step": 14330 + }, + { + "epoch": 0.553689331634426, + "grad_norm": 0.44783294200897217, + "learning_rate": 0.00016308995199299845, + "loss": 0.4745, + "step": 14340 + }, + { + "epoch": 0.5540754469284528, + "grad_norm": 1.0200363397598267, + "learning_rate": 0.00016306421097339667, + "loss": 0.1251, + "step": 14350 + }, + { + "epoch": 0.5544615622224797, + "grad_norm": 1.1761879920959473, + "learning_rate": 0.00016303846995379488, + "loss": 0.6837, + "step": 14360 + }, + { + "epoch": 0.5548476775165064, + "grad_norm": 1.8275704383850098, + "learning_rate": 0.0001630127289341931, + "loss": 0.3968, + "step": 14370 + }, + { + "epoch": 0.5552337928105332, + "grad_norm": 0.7219232320785522, + "learning_rate": 0.0001629869879145913, + "loss": 0.3278, + "step": 14380 + }, + { + "epoch": 0.55561990810456, + "grad_norm": 1.9161540269851685, + "learning_rate": 0.00016296124689498952, + "loss": 0.5071, + "step": 14390 + }, + { + "epoch": 0.5560060233985868, + "grad_norm": 2.4773502349853516, + "learning_rate": 0.00016293550587538773, + "loss": 0.3268, + "step": 14400 + }, + { + "epoch": 0.5563921386926136, + "grad_norm": 1.526877760887146, + "learning_rate": 0.00016290976485578595, + "loss": 0.284, + "step": 14410 + }, + { + "epoch": 0.5567782539866404, + "grad_norm": 2.082036018371582, + "learning_rate": 0.00016288402383618416, + "loss": 0.4194, + "step": 14420 + }, + { + "epoch": 0.5571643692806673, + "grad_norm": 4.1033477783203125, + "learning_rate": 0.00016285828281658237, + "loss": 0.329, + "step": 14430 + }, + { + "epoch": 0.557550484574694, + "grad_norm": 3.344879388809204, + "learning_rate": 0.0001628325417969806, + "loss": 0.3599, + "step": 14440 + }, + { + "epoch": 0.5579365998687208, + "grad_norm": 2.6200602054595947, + "learning_rate": 0.0001628068007773788, + "loss": 0.5152, + "step": 14450 + }, + { + "epoch": 0.5583227151627476, + "grad_norm": 1.05362868309021, + "learning_rate": 0.000162781059757777, + "loss": 0.4454, + "step": 14460 + }, + { + "epoch": 0.5587088304567744, + "grad_norm": 2.557406187057495, + "learning_rate": 0.00016275531873817523, + "loss": 0.3779, + "step": 14470 + }, + { + "epoch": 0.5590949457508012, + "grad_norm": 0.8478209376335144, + "learning_rate": 0.00016272957771857344, + "loss": 0.4289, + "step": 14480 + }, + { + "epoch": 0.559481061044828, + "grad_norm": 3.543574094772339, + "learning_rate": 0.00016270383669897165, + "loss": 0.37, + "step": 14490 + }, + { + "epoch": 0.5598671763388547, + "grad_norm": 0.21068768203258514, + "learning_rate": 0.00016267809567936987, + "loss": 0.2602, + "step": 14500 + }, + { + "epoch": 0.5602532916328816, + "grad_norm": 1.1703628301620483, + "learning_rate": 0.00016265235465976808, + "loss": 0.3684, + "step": 14510 + }, + { + "epoch": 0.5606394069269084, + "grad_norm": 1.4498575925827026, + "learning_rate": 0.0001626266136401663, + "loss": 0.4089, + "step": 14520 + }, + { + "epoch": 0.5610255222209352, + "grad_norm": 1.617297887802124, + "learning_rate": 0.0001626008726205645, + "loss": 0.2759, + "step": 14530 + }, + { + "epoch": 0.5614116375149619, + "grad_norm": 0.8708978891372681, + "learning_rate": 0.00016257513160096272, + "loss": 0.2523, + "step": 14540 + }, + { + "epoch": 0.5617977528089888, + "grad_norm": 1.098026156425476, + "learning_rate": 0.00016254939058136093, + "loss": 0.3507, + "step": 14550 + }, + { + "epoch": 0.5621838681030156, + "grad_norm": 3.0867936611175537, + "learning_rate": 0.00016252364956175915, + "loss": 0.3563, + "step": 14560 + }, + { + "epoch": 0.5625699833970423, + "grad_norm": 1.4829964637756348, + "learning_rate": 0.00016249790854215736, + "loss": 0.3586, + "step": 14570 + }, + { + "epoch": 0.5629560986910691, + "grad_norm": 4.029405117034912, + "learning_rate": 0.00016247216752255557, + "loss": 0.5198, + "step": 14580 + }, + { + "epoch": 0.563342213985096, + "grad_norm": 2.5473573207855225, + "learning_rate": 0.00016244642650295379, + "loss": 0.3818, + "step": 14590 + }, + { + "epoch": 0.5637283292791228, + "grad_norm": 1.387529730796814, + "learning_rate": 0.000162420685483352, + "loss": 0.3453, + "step": 14600 + }, + { + "epoch": 0.5641144445731495, + "grad_norm": 3.2525246143341064, + "learning_rate": 0.0001623949444637502, + "loss": 0.6096, + "step": 14610 + }, + { + "epoch": 0.5645005598671763, + "grad_norm": 1.23868989944458, + "learning_rate": 0.00016236920344414843, + "loss": 0.2785, + "step": 14620 + }, + { + "epoch": 0.5648866751612032, + "grad_norm": 1.763416051864624, + "learning_rate": 0.00016234346242454667, + "loss": 0.2313, + "step": 14630 + }, + { + "epoch": 0.5652727904552299, + "grad_norm": 2.6030027866363525, + "learning_rate": 0.00016231772140494485, + "loss": 0.5002, + "step": 14640 + }, + { + "epoch": 0.5656589057492567, + "grad_norm": 4.345195770263672, + "learning_rate": 0.00016229198038534307, + "loss": 0.3433, + "step": 14650 + }, + { + "epoch": 0.5660450210432835, + "grad_norm": 1.8660558462142944, + "learning_rate": 0.00016226623936574128, + "loss": 0.325, + "step": 14660 + }, + { + "epoch": 0.5664311363373103, + "grad_norm": 2.504354953765869, + "learning_rate": 0.0001622404983461395, + "loss": 0.3013, + "step": 14670 + }, + { + "epoch": 0.5668172516313371, + "grad_norm": 1.715135097503662, + "learning_rate": 0.0001622147573265377, + "loss": 0.2396, + "step": 14680 + }, + { + "epoch": 0.5672033669253639, + "grad_norm": 0.5195931792259216, + "learning_rate": 0.00016218901630693592, + "loss": 0.4992, + "step": 14690 + }, + { + "epoch": 0.5675894822193908, + "grad_norm": 1.076095461845398, + "learning_rate": 0.00016216327528733416, + "loss": 0.4488, + "step": 14700 + }, + { + "epoch": 0.5679755975134175, + "grad_norm": 0.42350637912750244, + "learning_rate": 0.00016213753426773235, + "loss": 0.2798, + "step": 14710 + }, + { + "epoch": 0.5683617128074443, + "grad_norm": 2.8514647483825684, + "learning_rate": 0.00016211179324813056, + "loss": 0.3108, + "step": 14720 + }, + { + "epoch": 0.5687478281014711, + "grad_norm": 1.4496532678604126, + "learning_rate": 0.00016208605222852877, + "loss": 0.4197, + "step": 14730 + }, + { + "epoch": 0.5691339433954979, + "grad_norm": 2.513998031616211, + "learning_rate": 0.00016206031120892699, + "loss": 0.4931, + "step": 14740 + }, + { + "epoch": 0.5695200586895247, + "grad_norm": 1.5905802249908447, + "learning_rate": 0.0001620345701893252, + "loss": 0.2175, + "step": 14750 + }, + { + "epoch": 0.5699061739835515, + "grad_norm": 0.4657856523990631, + "learning_rate": 0.0001620088291697234, + "loss": 0.4753, + "step": 14760 + }, + { + "epoch": 0.5702922892775782, + "grad_norm": 1.4188594818115234, + "learning_rate": 0.00016198308815012165, + "loss": 0.4849, + "step": 14770 + }, + { + "epoch": 0.5706784045716051, + "grad_norm": 0.6742203235626221, + "learning_rate": 0.00016195734713051984, + "loss": 0.3451, + "step": 14780 + }, + { + "epoch": 0.5710645198656319, + "grad_norm": 1.521262526512146, + "learning_rate": 0.00016193160611091805, + "loss": 0.6253, + "step": 14790 + }, + { + "epoch": 0.5714506351596587, + "grad_norm": 0.8657771348953247, + "learning_rate": 0.00016190586509131627, + "loss": 0.3664, + "step": 14800 + }, + { + "epoch": 0.5718367504536854, + "grad_norm": 1.0131505727767944, + "learning_rate": 0.00016188012407171448, + "loss": 0.3395, + "step": 14810 + }, + { + "epoch": 0.5722228657477123, + "grad_norm": 0.8506319522857666, + "learning_rate": 0.00016185438305211272, + "loss": 0.2769, + "step": 14820 + }, + { + "epoch": 0.5726089810417391, + "grad_norm": 3.1080141067504883, + "learning_rate": 0.0001618286420325109, + "loss": 0.3185, + "step": 14830 + }, + { + "epoch": 0.5729950963357658, + "grad_norm": 0.8805003762245178, + "learning_rate": 0.00016180290101290915, + "loss": 0.3089, + "step": 14840 + }, + { + "epoch": 0.5733812116297926, + "grad_norm": 3.9470136165618896, + "learning_rate": 0.00016177715999330733, + "loss": 0.4552, + "step": 14850 + }, + { + "epoch": 0.5737673269238195, + "grad_norm": 1.10677969455719, + "learning_rate": 0.00016175141897370557, + "loss": 0.4624, + "step": 14860 + }, + { + "epoch": 0.5741534422178463, + "grad_norm": 1.185539960861206, + "learning_rate": 0.00016172567795410376, + "loss": 0.3144, + "step": 14870 + }, + { + "epoch": 0.574539557511873, + "grad_norm": 2.484386920928955, + "learning_rate": 0.00016169993693450197, + "loss": 0.3409, + "step": 14880 + }, + { + "epoch": 0.5749256728058998, + "grad_norm": 2.1621437072753906, + "learning_rate": 0.0001616741959149002, + "loss": 0.4421, + "step": 14890 + }, + { + "epoch": 0.5753117880999267, + "grad_norm": 1.3359025716781616, + "learning_rate": 0.0001616484548952984, + "loss": 0.2826, + "step": 14900 + }, + { + "epoch": 0.5756979033939534, + "grad_norm": 2.828157901763916, + "learning_rate": 0.00016162271387569664, + "loss": 0.4615, + "step": 14910 + }, + { + "epoch": 0.5760840186879802, + "grad_norm": 2.0543019771575928, + "learning_rate": 0.00016159697285609483, + "loss": 0.4307, + "step": 14920 + }, + { + "epoch": 0.5764701339820071, + "grad_norm": 0.3085225522518158, + "learning_rate": 0.00016157123183649307, + "loss": 0.317, + "step": 14930 + }, + { + "epoch": 0.5768562492760339, + "grad_norm": 1.459349274635315, + "learning_rate": 0.00016154549081689125, + "loss": 0.3928, + "step": 14940 + }, + { + "epoch": 0.5772423645700606, + "grad_norm": 0.6684612035751343, + "learning_rate": 0.00016151974979728947, + "loss": 0.4129, + "step": 14950 + }, + { + "epoch": 0.5776284798640874, + "grad_norm": 2.428311586380005, + "learning_rate": 0.0001614940087776877, + "loss": 0.4163, + "step": 14960 + }, + { + "epoch": 0.5780145951581143, + "grad_norm": 1.8885403871536255, + "learning_rate": 0.0001614682677580859, + "loss": 0.4311, + "step": 14970 + }, + { + "epoch": 0.578400710452141, + "grad_norm": 3.9598031044006348, + "learning_rate": 0.00016144252673848413, + "loss": 0.3103, + "step": 14980 + }, + { + "epoch": 0.5787868257461678, + "grad_norm": 1.872383713722229, + "learning_rate": 0.00016141678571888232, + "loss": 0.3592, + "step": 14990 + }, + { + "epoch": 0.5791729410401946, + "grad_norm": 1.023526668548584, + "learning_rate": 0.00016139104469928056, + "loss": 0.4185, + "step": 15000 + }, + { + "epoch": 0.5795590563342214, + "grad_norm": 1.5721429586410522, + "learning_rate": 0.00016136530367967877, + "loss": 0.4017, + "step": 15010 + }, + { + "epoch": 0.5799451716282482, + "grad_norm": 3.502350091934204, + "learning_rate": 0.00016133956266007696, + "loss": 0.397, + "step": 15020 + }, + { + "epoch": 0.580331286922275, + "grad_norm": 2.415985345840454, + "learning_rate": 0.0001613138216404752, + "loss": 0.4041, + "step": 15030 + }, + { + "epoch": 0.5807174022163017, + "grad_norm": 0.5441868901252747, + "learning_rate": 0.00016128808062087339, + "loss": 0.2395, + "step": 15040 + }, + { + "epoch": 0.5811035175103286, + "grad_norm": 2.453216552734375, + "learning_rate": 0.00016126233960127163, + "loss": 0.1586, + "step": 15050 + }, + { + "epoch": 0.5814896328043554, + "grad_norm": 3.108646869659424, + "learning_rate": 0.0001612365985816698, + "loss": 0.3996, + "step": 15060 + }, + { + "epoch": 0.5818757480983822, + "grad_norm": 0.7707905173301697, + "learning_rate": 0.00016121085756206805, + "loss": 0.1756, + "step": 15070 + }, + { + "epoch": 0.5822618633924089, + "grad_norm": 0.42857447266578674, + "learning_rate": 0.00016118511654246627, + "loss": 0.258, + "step": 15080 + }, + { + "epoch": 0.5826479786864358, + "grad_norm": 0.7148373126983643, + "learning_rate": 0.00016115937552286445, + "loss": 0.3755, + "step": 15090 + }, + { + "epoch": 0.5830340939804626, + "grad_norm": 0.04789021611213684, + "learning_rate": 0.0001611336345032627, + "loss": 0.2087, + "step": 15100 + }, + { + "epoch": 0.5834202092744893, + "grad_norm": 5.012516975402832, + "learning_rate": 0.00016110789348366088, + "loss": 0.5406, + "step": 15110 + }, + { + "epoch": 0.5838063245685161, + "grad_norm": 1.4139299392700195, + "learning_rate": 0.00016108215246405912, + "loss": 0.407, + "step": 15120 + }, + { + "epoch": 0.584192439862543, + "grad_norm": 0.8637074828147888, + "learning_rate": 0.00016105641144445733, + "loss": 0.2987, + "step": 15130 + }, + { + "epoch": 0.5845785551565698, + "grad_norm": 0.9556403160095215, + "learning_rate": 0.00016103067042485555, + "loss": 0.4319, + "step": 15140 + }, + { + "epoch": 0.5849646704505965, + "grad_norm": 2.071455955505371, + "learning_rate": 0.00016100492940525376, + "loss": 0.4824, + "step": 15150 + }, + { + "epoch": 0.5853507857446233, + "grad_norm": 3.8130764961242676, + "learning_rate": 0.00016097918838565195, + "loss": 0.4749, + "step": 15160 + }, + { + "epoch": 0.5857369010386502, + "grad_norm": 1.290738582611084, + "learning_rate": 0.00016095344736605019, + "loss": 0.4486, + "step": 15170 + }, + { + "epoch": 0.5861230163326769, + "grad_norm": 0.9964671730995178, + "learning_rate": 0.00016092770634644837, + "loss": 0.1514, + "step": 15180 + }, + { + "epoch": 0.5865091316267037, + "grad_norm": 0.5267524123191833, + "learning_rate": 0.0001609019653268466, + "loss": 0.2298, + "step": 15190 + }, + { + "epoch": 0.5868952469207306, + "grad_norm": 2.028960704803467, + "learning_rate": 0.00016087622430724483, + "loss": 0.2925, + "step": 15200 + }, + { + "epoch": 0.5872813622147574, + "grad_norm": 0.8423904776573181, + "learning_rate": 0.00016085048328764304, + "loss": 0.4221, + "step": 15210 + }, + { + "epoch": 0.5876674775087841, + "grad_norm": 1.9663047790527344, + "learning_rate": 0.00016082474226804125, + "loss": 0.3595, + "step": 15220 + }, + { + "epoch": 0.5880535928028109, + "grad_norm": 1.2221906185150146, + "learning_rate": 0.00016079900124843944, + "loss": 0.3174, + "step": 15230 + }, + { + "epoch": 0.5884397080968378, + "grad_norm": 2.138437032699585, + "learning_rate": 0.00016077326022883768, + "loss": 0.2993, + "step": 15240 + }, + { + "epoch": 0.5888258233908645, + "grad_norm": 1.8036224842071533, + "learning_rate": 0.00016074751920923586, + "loss": 0.3897, + "step": 15250 + }, + { + "epoch": 0.5892119386848913, + "grad_norm": 2.3102879524230957, + "learning_rate": 0.0001607217781896341, + "loss": 0.3713, + "step": 15260 + }, + { + "epoch": 0.5895980539789181, + "grad_norm": 1.40048348903656, + "learning_rate": 0.00016069603717003232, + "loss": 0.2701, + "step": 15270 + }, + { + "epoch": 0.589984169272945, + "grad_norm": 1.0581787824630737, + "learning_rate": 0.00016067029615043053, + "loss": 0.2529, + "step": 15280 + }, + { + "epoch": 0.5903702845669717, + "grad_norm": 0.668211042881012, + "learning_rate": 0.00016064455513082875, + "loss": 0.221, + "step": 15290 + }, + { + "epoch": 0.5907563998609985, + "grad_norm": 0.7950372099876404, + "learning_rate": 0.00016061881411122696, + "loss": 0.2405, + "step": 15300 + }, + { + "epoch": 0.5911425151550252, + "grad_norm": 1.8531723022460938, + "learning_rate": 0.00016059307309162517, + "loss": 0.3423, + "step": 15310 + }, + { + "epoch": 0.5915286304490521, + "grad_norm": 0.2071121335029602, + "learning_rate": 0.00016056733207202339, + "loss": 0.2923, + "step": 15320 + }, + { + "epoch": 0.5919147457430789, + "grad_norm": 2.4298369884490967, + "learning_rate": 0.0001605415910524216, + "loss": 0.531, + "step": 15330 + }, + { + "epoch": 0.5923008610371057, + "grad_norm": 3.2297933101654053, + "learning_rate": 0.0001605158500328198, + "loss": 0.2563, + "step": 15340 + }, + { + "epoch": 0.5926869763311324, + "grad_norm": 1.533225178718567, + "learning_rate": 0.00016049010901321803, + "loss": 0.2712, + "step": 15350 + }, + { + "epoch": 0.5930730916251593, + "grad_norm": 3.6168954372406006, + "learning_rate": 0.00016046436799361624, + "loss": 0.6428, + "step": 15360 + }, + { + "epoch": 0.5934592069191861, + "grad_norm": 0.8912101984024048, + "learning_rate": 0.00016043862697401445, + "loss": 0.2882, + "step": 15370 + }, + { + "epoch": 0.5938453222132128, + "grad_norm": 0.6782923936843872, + "learning_rate": 0.00016041288595441267, + "loss": 0.3451, + "step": 15380 + }, + { + "epoch": 0.5942314375072396, + "grad_norm": 2.7575254440307617, + "learning_rate": 0.00016038714493481088, + "loss": 0.272, + "step": 15390 + }, + { + "epoch": 0.5946175528012665, + "grad_norm": 1.8348017930984497, + "learning_rate": 0.0001603614039152091, + "loss": 0.286, + "step": 15400 + }, + { + "epoch": 0.5950036680952933, + "grad_norm": 3.1459157466888428, + "learning_rate": 0.0001603356628956073, + "loss": 0.2986, + "step": 15410 + }, + { + "epoch": 0.59538978338932, + "grad_norm": 2.0769810676574707, + "learning_rate": 0.00016030992187600552, + "loss": 0.5512, + "step": 15420 + }, + { + "epoch": 0.5957758986833468, + "grad_norm": 0.5503840446472168, + "learning_rate": 0.00016028418085640373, + "loss": 0.4117, + "step": 15430 + }, + { + "epoch": 0.5961620139773737, + "grad_norm": 1.9759409427642822, + "learning_rate": 0.00016025843983680194, + "loss": 0.3619, + "step": 15440 + }, + { + "epoch": 0.5965481292714004, + "grad_norm": 2.2451424598693848, + "learning_rate": 0.00016023269881720016, + "loss": 0.2798, + "step": 15450 + }, + { + "epoch": 0.5969342445654272, + "grad_norm": 1.8537431955337524, + "learning_rate": 0.00016020695779759837, + "loss": 0.3739, + "step": 15460 + }, + { + "epoch": 0.5973203598594541, + "grad_norm": 1.7667044401168823, + "learning_rate": 0.00016018121677799658, + "loss": 0.3694, + "step": 15470 + }, + { + "epoch": 0.5977064751534809, + "grad_norm": 0.8955661654472351, + "learning_rate": 0.0001601554757583948, + "loss": 0.2036, + "step": 15480 + }, + { + "epoch": 0.5980925904475076, + "grad_norm": 0.9526143074035645, + "learning_rate": 0.000160129734738793, + "loss": 0.3728, + "step": 15490 + }, + { + "epoch": 0.5984787057415344, + "grad_norm": 0.5816594958305359, + "learning_rate": 0.00016010399371919122, + "loss": 0.3263, + "step": 15500 + }, + { + "epoch": 0.5988648210355613, + "grad_norm": 0.6841669678688049, + "learning_rate": 0.00016007825269958944, + "loss": 0.3252, + "step": 15510 + }, + { + "epoch": 0.599250936329588, + "grad_norm": 0.9375134706497192, + "learning_rate": 0.00016005251167998765, + "loss": 0.2207, + "step": 15520 + }, + { + "epoch": 0.5996370516236148, + "grad_norm": 0.39535248279571533, + "learning_rate": 0.00016002677066038586, + "loss": 0.3566, + "step": 15530 + }, + { + "epoch": 0.6000231669176416, + "grad_norm": 0.5440202951431274, + "learning_rate": 0.00016000102964078408, + "loss": 0.2886, + "step": 15540 + }, + { + "epoch": 0.6004092822116684, + "grad_norm": 0.45111024379730225, + "learning_rate": 0.0001599752886211823, + "loss": 0.2697, + "step": 15550 + }, + { + "epoch": 0.6007953975056952, + "grad_norm": 2.372063398361206, + "learning_rate": 0.0001599495476015805, + "loss": 0.4716, + "step": 15560 + }, + { + "epoch": 0.601181512799722, + "grad_norm": 0.5841318368911743, + "learning_rate": 0.00015992380658197872, + "loss": 0.5611, + "step": 15570 + }, + { + "epoch": 0.6015676280937488, + "grad_norm": 0.323010116815567, + "learning_rate": 0.00015989806556237693, + "loss": 0.2733, + "step": 15580 + }, + { + "epoch": 0.6019537433877756, + "grad_norm": 1.4498323202133179, + "learning_rate": 0.00015987232454277514, + "loss": 0.4063, + "step": 15590 + }, + { + "epoch": 0.6023398586818024, + "grad_norm": 0.47180086374282837, + "learning_rate": 0.00015984658352317336, + "loss": 0.3322, + "step": 15600 + }, + { + "epoch": 0.6027259739758292, + "grad_norm": 1.2303547859191895, + "learning_rate": 0.00015982084250357157, + "loss": 0.2985, + "step": 15610 + }, + { + "epoch": 0.6031120892698559, + "grad_norm": 2.5056209564208984, + "learning_rate": 0.00015979510148396978, + "loss": 0.4903, + "step": 15620 + }, + { + "epoch": 0.6034982045638828, + "grad_norm": 1.344814419746399, + "learning_rate": 0.00015976936046436802, + "loss": 0.3806, + "step": 15630 + }, + { + "epoch": 0.6038843198579096, + "grad_norm": 3.2931411266326904, + "learning_rate": 0.0001597436194447662, + "loss": 0.3291, + "step": 15640 + }, + { + "epoch": 0.6042704351519363, + "grad_norm": 0.3108818829059601, + "learning_rate": 0.00015971787842516442, + "loss": 0.2866, + "step": 15650 + }, + { + "epoch": 0.6046565504459631, + "grad_norm": 2.6437489986419678, + "learning_rate": 0.00015969213740556264, + "loss": 0.2853, + "step": 15660 + }, + { + "epoch": 0.60504266573999, + "grad_norm": 0.6996239423751831, + "learning_rate": 0.00015966639638596085, + "loss": 0.2947, + "step": 15670 + }, + { + "epoch": 0.6054287810340168, + "grad_norm": 0.9377492070198059, + "learning_rate": 0.00015964065536635906, + "loss": 0.5502, + "step": 15680 + }, + { + "epoch": 0.6058148963280435, + "grad_norm": 0.3067781627178192, + "learning_rate": 0.00015961491434675728, + "loss": 0.2916, + "step": 15690 + }, + { + "epoch": 0.6062010116220703, + "grad_norm": 1.6191383600234985, + "learning_rate": 0.00015958917332715552, + "loss": 0.2536, + "step": 15700 + }, + { + "epoch": 0.6065871269160972, + "grad_norm": 0.5139639973640442, + "learning_rate": 0.0001595634323075537, + "loss": 0.2809, + "step": 15710 + }, + { + "epoch": 0.6069732422101239, + "grad_norm": 1.6476198434829712, + "learning_rate": 0.00015953769128795192, + "loss": 0.2859, + "step": 15720 + }, + { + "epoch": 0.6073593575041507, + "grad_norm": 3.895970106124878, + "learning_rate": 0.00015951195026835013, + "loss": 0.5254, + "step": 15730 + }, + { + "epoch": 0.6077454727981776, + "grad_norm": 1.1022089719772339, + "learning_rate": 0.00015948620924874834, + "loss": 0.3254, + "step": 15740 + }, + { + "epoch": 0.6081315880922044, + "grad_norm": 1.3811163902282715, + "learning_rate": 0.00015946046822914656, + "loss": 0.3259, + "step": 15750 + }, + { + "epoch": 0.6085177033862311, + "grad_norm": 1.0810881853103638, + "learning_rate": 0.00015943472720954477, + "loss": 0.3689, + "step": 15760 + }, + { + "epoch": 0.6089038186802579, + "grad_norm": 2.1972954273223877, + "learning_rate": 0.000159408986189943, + "loss": 0.3255, + "step": 15770 + }, + { + "epoch": 0.6092899339742848, + "grad_norm": 3.945215940475464, + "learning_rate": 0.0001593832451703412, + "loss": 0.4317, + "step": 15780 + }, + { + "epoch": 0.6096760492683115, + "grad_norm": 0.5246737599372864, + "learning_rate": 0.0001593575041507394, + "loss": 0.5895, + "step": 15790 + }, + { + "epoch": 0.6100621645623383, + "grad_norm": 0.0722908228635788, + "learning_rate": 0.00015933176313113762, + "loss": 0.2078, + "step": 15800 + }, + { + "epoch": 0.6104482798563651, + "grad_norm": 1.888529658317566, + "learning_rate": 0.00015930602211153584, + "loss": 0.3629, + "step": 15810 + }, + { + "epoch": 0.610834395150392, + "grad_norm": 0.7190912365913391, + "learning_rate": 0.00015928028109193408, + "loss": 0.4715, + "step": 15820 + }, + { + "epoch": 0.6112205104444187, + "grad_norm": 1.1624583005905151, + "learning_rate": 0.00015925454007233226, + "loss": 0.3678, + "step": 15830 + }, + { + "epoch": 0.6116066257384455, + "grad_norm": 1.4570399522781372, + "learning_rate": 0.0001592287990527305, + "loss": 0.3008, + "step": 15840 + }, + { + "epoch": 0.6119927410324723, + "grad_norm": 0.45386505126953125, + "learning_rate": 0.0001592030580331287, + "loss": 0.2919, + "step": 15850 + }, + { + "epoch": 0.6123788563264991, + "grad_norm": 1.9031087160110474, + "learning_rate": 0.0001591773170135269, + "loss": 0.4336, + "step": 15860 + }, + { + "epoch": 0.6127649716205259, + "grad_norm": 0.9156181216239929, + "learning_rate": 0.00015915157599392512, + "loss": 0.2451, + "step": 15870 + }, + { + "epoch": 0.6131510869145527, + "grad_norm": 1.8031158447265625, + "learning_rate": 0.00015912583497432333, + "loss": 0.259, + "step": 15880 + }, + { + "epoch": 0.6135372022085794, + "grad_norm": 1.4521692991256714, + "learning_rate": 0.00015910009395472157, + "loss": 0.279, + "step": 15890 + }, + { + "epoch": 0.6139233175026063, + "grad_norm": 1.5064165592193604, + "learning_rate": 0.00015907435293511976, + "loss": 0.34, + "step": 15900 + }, + { + "epoch": 0.6143094327966331, + "grad_norm": 0.588637113571167, + "learning_rate": 0.000159048611915518, + "loss": 0.4886, + "step": 15910 + }, + { + "epoch": 0.6146955480906598, + "grad_norm": 0.3159797191619873, + "learning_rate": 0.00015902287089591618, + "loss": 0.3265, + "step": 15920 + }, + { + "epoch": 0.6150816633846866, + "grad_norm": 3.3988165855407715, + "learning_rate": 0.0001589971298763144, + "loss": 0.4917, + "step": 15930 + }, + { + "epoch": 0.6154677786787135, + "grad_norm": 0.5139709711074829, + "learning_rate": 0.00015897138885671264, + "loss": 0.2175, + "step": 15940 + }, + { + "epoch": 0.6158538939727403, + "grad_norm": 3.6877944469451904, + "learning_rate": 0.00015894564783711082, + "loss": 0.4674, + "step": 15950 + }, + { + "epoch": 0.616240009266767, + "grad_norm": 1.6468040943145752, + "learning_rate": 0.00015891990681750906, + "loss": 0.4375, + "step": 15960 + }, + { + "epoch": 0.6166261245607938, + "grad_norm": 0.47542962431907654, + "learning_rate": 0.00015889416579790725, + "loss": 0.3593, + "step": 15970 + }, + { + "epoch": 0.6170122398548207, + "grad_norm": 2.210597038269043, + "learning_rate": 0.0001588684247783055, + "loss": 0.3356, + "step": 15980 + }, + { + "epoch": 0.6173983551488474, + "grad_norm": 2.0030908584594727, + "learning_rate": 0.00015884268375870368, + "loss": 0.3367, + "step": 15990 + }, + { + "epoch": 0.6177844704428742, + "grad_norm": 3.438887119293213, + "learning_rate": 0.0001588169427391019, + "loss": 0.4583, + "step": 16000 + }, + { + "epoch": 0.6181705857369011, + "grad_norm": 2.88147234916687, + "learning_rate": 0.00015879120171950013, + "loss": 0.3837, + "step": 16010 + }, + { + "epoch": 0.6185567010309279, + "grad_norm": 0.9327366948127747, + "learning_rate": 0.00015876546069989832, + "loss": 0.236, + "step": 16020 + }, + { + "epoch": 0.6189428163249546, + "grad_norm": 2.205355405807495, + "learning_rate": 0.00015873971968029656, + "loss": 0.2957, + "step": 16030 + }, + { + "epoch": 0.6193289316189814, + "grad_norm": 2.3861300945281982, + "learning_rate": 0.00015871397866069474, + "loss": 0.427, + "step": 16040 + }, + { + "epoch": 0.6197150469130083, + "grad_norm": 0.3276061415672302, + "learning_rate": 0.00015868823764109298, + "loss": 0.2301, + "step": 16050 + }, + { + "epoch": 0.620101162207035, + "grad_norm": 0.6325292587280273, + "learning_rate": 0.00015866249662149117, + "loss": 0.3915, + "step": 16060 + }, + { + "epoch": 0.6204872775010618, + "grad_norm": 1.1546003818511963, + "learning_rate": 0.0001586367556018894, + "loss": 0.4081, + "step": 16070 + }, + { + "epoch": 0.6208733927950886, + "grad_norm": 1.7260868549346924, + "learning_rate": 0.00015861101458228762, + "loss": 0.4642, + "step": 16080 + }, + { + "epoch": 0.6212595080891155, + "grad_norm": 0.30561816692352295, + "learning_rate": 0.0001585852735626858, + "loss": 0.2065, + "step": 16090 + }, + { + "epoch": 0.6216456233831422, + "grad_norm": 2.440007448196411, + "learning_rate": 0.00015855953254308405, + "loss": 0.3851, + "step": 16100 + }, + { + "epoch": 0.622031738677169, + "grad_norm": 4.06764554977417, + "learning_rate": 0.00015853379152348224, + "loss": 0.3796, + "step": 16110 + }, + { + "epoch": 0.6224178539711958, + "grad_norm": 4.970936298370361, + "learning_rate": 0.00015850805050388048, + "loss": 0.3393, + "step": 16120 + }, + { + "epoch": 0.6228039692652226, + "grad_norm": 2.0721356868743896, + "learning_rate": 0.0001584823094842787, + "loss": 0.2991, + "step": 16130 + }, + { + "epoch": 0.6231900845592494, + "grad_norm": 1.8130602836608887, + "learning_rate": 0.0001584565684646769, + "loss": 0.3132, + "step": 16140 + }, + { + "epoch": 0.6235761998532762, + "grad_norm": 1.1093038320541382, + "learning_rate": 0.00015843082744507512, + "loss": 0.3144, + "step": 16150 + }, + { + "epoch": 0.6239623151473029, + "grad_norm": 2.227937936782837, + "learning_rate": 0.0001584050864254733, + "loss": 0.4955, + "step": 16160 + }, + { + "epoch": 0.6243484304413298, + "grad_norm": 1.2541782855987549, + "learning_rate": 0.00015837934540587154, + "loss": 0.225, + "step": 16170 + }, + { + "epoch": 0.6247345457353566, + "grad_norm": 1.8113441467285156, + "learning_rate": 0.00015835360438626973, + "loss": 0.5393, + "step": 16180 + }, + { + "epoch": 0.6251206610293834, + "grad_norm": 2.198061466217041, + "learning_rate": 0.00015832786336666797, + "loss": 0.2225, + "step": 16190 + }, + { + "epoch": 0.6255067763234101, + "grad_norm": 2.4241714477539062, + "learning_rate": 0.00015830212234706618, + "loss": 0.3294, + "step": 16200 + }, + { + "epoch": 0.625892891617437, + "grad_norm": 3.0632903575897217, + "learning_rate": 0.0001582763813274644, + "loss": 0.3776, + "step": 16210 + }, + { + "epoch": 0.6262790069114638, + "grad_norm": 1.9433149099349976, + "learning_rate": 0.0001582506403078626, + "loss": 0.2964, + "step": 16220 + }, + { + "epoch": 0.6266651222054905, + "grad_norm": 1.6328935623168945, + "learning_rate": 0.0001582248992882608, + "loss": 0.2169, + "step": 16230 + }, + { + "epoch": 0.6270512374995174, + "grad_norm": 0.926477313041687, + "learning_rate": 0.00015819915826865904, + "loss": 0.271, + "step": 16240 + }, + { + "epoch": 0.6274373527935442, + "grad_norm": 3.4526686668395996, + "learning_rate": 0.00015817341724905722, + "loss": 0.1836, + "step": 16250 + }, + { + "epoch": 0.627823468087571, + "grad_norm": 0.8149943947792053, + "learning_rate": 0.00015814767622945546, + "loss": 0.2494, + "step": 16260 + }, + { + "epoch": 0.6282095833815977, + "grad_norm": 0.8609616756439209, + "learning_rate": 0.00015812193520985368, + "loss": 0.279, + "step": 16270 + }, + { + "epoch": 0.6285956986756246, + "grad_norm": 0.9432594180107117, + "learning_rate": 0.0001580961941902519, + "loss": 0.3485, + "step": 16280 + }, + { + "epoch": 0.6289818139696514, + "grad_norm": 2.8844246864318848, + "learning_rate": 0.0001580704531706501, + "loss": 0.6163, + "step": 16290 + }, + { + "epoch": 0.6293679292636781, + "grad_norm": 3.4325804710388184, + "learning_rate": 0.0001580447121510483, + "loss": 0.3392, + "step": 16300 + }, + { + "epoch": 0.6297540445577049, + "grad_norm": 0.5320155620574951, + "learning_rate": 0.00015801897113144653, + "loss": 0.4664, + "step": 16310 + }, + { + "epoch": 0.6301401598517318, + "grad_norm": 1.075914740562439, + "learning_rate": 0.00015799323011184474, + "loss": 0.3023, + "step": 16320 + }, + { + "epoch": 0.6305262751457585, + "grad_norm": 1.3115136623382568, + "learning_rate": 0.00015796748909224296, + "loss": 0.1741, + "step": 16330 + }, + { + "epoch": 0.6309123904397853, + "grad_norm": 0.18413056433200836, + "learning_rate": 0.00015794174807264117, + "loss": 0.1966, + "step": 16340 + }, + { + "epoch": 0.6312985057338121, + "grad_norm": 1.8707069158554077, + "learning_rate": 0.00015791600705303938, + "loss": 0.3175, + "step": 16350 + }, + { + "epoch": 0.631684621027839, + "grad_norm": 2.2436699867248535, + "learning_rate": 0.0001578902660334376, + "loss": 0.2619, + "step": 16360 + }, + { + "epoch": 0.6320707363218657, + "grad_norm": 2.6100945472717285, + "learning_rate": 0.00015786452501383578, + "loss": 0.2874, + "step": 16370 + }, + { + "epoch": 0.6324568516158925, + "grad_norm": 1.455538034439087, + "learning_rate": 0.00015783878399423402, + "loss": 0.3956, + "step": 16380 + }, + { + "epoch": 0.6328429669099193, + "grad_norm": 1.1950361728668213, + "learning_rate": 0.00015781304297463224, + "loss": 0.3406, + "step": 16390 + }, + { + "epoch": 0.6332290822039461, + "grad_norm": 0.6905789971351624, + "learning_rate": 0.00015778730195503045, + "loss": 0.2788, + "step": 16400 + }, + { + "epoch": 0.6336151974979729, + "grad_norm": 1.8803350925445557, + "learning_rate": 0.00015776156093542866, + "loss": 0.5509, + "step": 16410 + }, + { + "epoch": 0.6340013127919997, + "grad_norm": 4.088913440704346, + "learning_rate": 0.00015773581991582688, + "loss": 0.5238, + "step": 16420 + }, + { + "epoch": 0.6343874280860264, + "grad_norm": 2.9464988708496094, + "learning_rate": 0.0001577100788962251, + "loss": 0.4721, + "step": 16430 + }, + { + "epoch": 0.6347735433800533, + "grad_norm": 2.005481719970703, + "learning_rate": 0.0001576843378766233, + "loss": 0.323, + "step": 16440 + }, + { + "epoch": 0.6351596586740801, + "grad_norm": 0.1693512350320816, + "learning_rate": 0.00015765859685702152, + "loss": 0.3459, + "step": 16450 + }, + { + "epoch": 0.6355457739681069, + "grad_norm": 1.6552183628082275, + "learning_rate": 0.00015763285583741973, + "loss": 0.4299, + "step": 16460 + }, + { + "epoch": 0.6359318892621336, + "grad_norm": 0.8498923182487488, + "learning_rate": 0.00015760711481781794, + "loss": 0.3665, + "step": 16470 + }, + { + "epoch": 0.6363180045561605, + "grad_norm": 1.098840594291687, + "learning_rate": 0.00015758137379821616, + "loss": 0.318, + "step": 16480 + }, + { + "epoch": 0.6367041198501873, + "grad_norm": 2.69606876373291, + "learning_rate": 0.00015755563277861437, + "loss": 0.3566, + "step": 16490 + }, + { + "epoch": 0.637090235144214, + "grad_norm": 1.4099249839782715, + "learning_rate": 0.00015752989175901258, + "loss": 0.2658, + "step": 16500 + }, + { + "epoch": 0.6374763504382409, + "grad_norm": 0.10336513817310333, + "learning_rate": 0.0001575041507394108, + "loss": 0.4608, + "step": 16510 + }, + { + "epoch": 0.6378624657322677, + "grad_norm": 2.224609136581421, + "learning_rate": 0.000157478409719809, + "loss": 0.2875, + "step": 16520 + }, + { + "epoch": 0.6382485810262944, + "grad_norm": 1.3182893991470337, + "learning_rate": 0.00015745266870020722, + "loss": 0.2972, + "step": 16530 + }, + { + "epoch": 0.6386346963203212, + "grad_norm": 0.8028392195701599, + "learning_rate": 0.00015742692768060544, + "loss": 0.3553, + "step": 16540 + }, + { + "epoch": 0.6390208116143481, + "grad_norm": 2.6714046001434326, + "learning_rate": 0.00015740118666100365, + "loss": 0.3965, + "step": 16550 + }, + { + "epoch": 0.6394069269083749, + "grad_norm": 0.6173273921012878, + "learning_rate": 0.00015737544564140186, + "loss": 0.4278, + "step": 16560 + }, + { + "epoch": 0.6397930422024016, + "grad_norm": 0.9547831416130066, + "learning_rate": 0.00015734970462180008, + "loss": 0.3854, + "step": 16570 + }, + { + "epoch": 0.6401791574964284, + "grad_norm": 1.1336010694503784, + "learning_rate": 0.0001573239636021983, + "loss": 0.3505, + "step": 16580 + }, + { + "epoch": 0.6405652727904553, + "grad_norm": 4.911902904510498, + "learning_rate": 0.0001572982225825965, + "loss": 0.6624, + "step": 16590 + }, + { + "epoch": 0.640951388084482, + "grad_norm": 2.472303628921509, + "learning_rate": 0.00015727248156299472, + "loss": 0.4479, + "step": 16600 + }, + { + "epoch": 0.6413375033785088, + "grad_norm": 1.7077019214630127, + "learning_rate": 0.00015724674054339293, + "loss": 0.3985, + "step": 16610 + }, + { + "epoch": 0.6417236186725356, + "grad_norm": 4.096541404724121, + "learning_rate": 0.00015722099952379114, + "loss": 0.4763, + "step": 16620 + }, + { + "epoch": 0.6421097339665625, + "grad_norm": 1.5275769233703613, + "learning_rate": 0.00015719525850418936, + "loss": 0.3792, + "step": 16630 + }, + { + "epoch": 0.6424958492605892, + "grad_norm": 0.33548179268836975, + "learning_rate": 0.00015716951748458757, + "loss": 0.3276, + "step": 16640 + }, + { + "epoch": 0.642881964554616, + "grad_norm": 4.142831325531006, + "learning_rate": 0.00015714377646498578, + "loss": 0.4906, + "step": 16650 + }, + { + "epoch": 0.6432680798486428, + "grad_norm": 1.5129270553588867, + "learning_rate": 0.000157118035445384, + "loss": 0.3077, + "step": 16660 + }, + { + "epoch": 0.6436541951426696, + "grad_norm": 2.0287647247314453, + "learning_rate": 0.0001570922944257822, + "loss": 0.43, + "step": 16670 + }, + { + "epoch": 0.6440403104366964, + "grad_norm": 2.1278579235076904, + "learning_rate": 0.00015706655340618042, + "loss": 0.3822, + "step": 16680 + }, + { + "epoch": 0.6444264257307232, + "grad_norm": 0.7621383666992188, + "learning_rate": 0.00015704081238657864, + "loss": 0.2556, + "step": 16690 + }, + { + "epoch": 0.6448125410247499, + "grad_norm": 1.888422966003418, + "learning_rate": 0.00015701507136697685, + "loss": 0.4687, + "step": 16700 + }, + { + "epoch": 0.6451986563187768, + "grad_norm": 2.09405517578125, + "learning_rate": 0.00015698933034737506, + "loss": 0.4447, + "step": 16710 + }, + { + "epoch": 0.6455847716128036, + "grad_norm": 0.8639073967933655, + "learning_rate": 0.00015696358932777328, + "loss": 0.224, + "step": 16720 + }, + { + "epoch": 0.6459708869068304, + "grad_norm": 1.1472411155700684, + "learning_rate": 0.0001569378483081715, + "loss": 0.315, + "step": 16730 + }, + { + "epoch": 0.6463570022008571, + "grad_norm": 2.4987192153930664, + "learning_rate": 0.0001569121072885697, + "loss": 0.3707, + "step": 16740 + }, + { + "epoch": 0.646743117494884, + "grad_norm": 1.661458134651184, + "learning_rate": 0.00015688636626896792, + "loss": 0.4027, + "step": 16750 + }, + { + "epoch": 0.6471292327889108, + "grad_norm": 1.5494720935821533, + "learning_rate": 0.00015686062524936613, + "loss": 0.3737, + "step": 16760 + }, + { + "epoch": 0.6475153480829375, + "grad_norm": 0.3376433253288269, + "learning_rate": 0.00015683488422976437, + "loss": 0.3468, + "step": 16770 + }, + { + "epoch": 0.6479014633769644, + "grad_norm": 0.6496528387069702, + "learning_rate": 0.00015680914321016256, + "loss": 0.2857, + "step": 16780 + }, + { + "epoch": 0.6482875786709912, + "grad_norm": 1.957741618156433, + "learning_rate": 0.00015678340219056077, + "loss": 0.5089, + "step": 16790 + }, + { + "epoch": 0.648673693965018, + "grad_norm": 3.980466842651367, + "learning_rate": 0.00015675766117095898, + "loss": 0.3558, + "step": 16800 + }, + { + "epoch": 0.6490598092590447, + "grad_norm": 3.2516696453094482, + "learning_rate": 0.0001567319201513572, + "loss": 0.6214, + "step": 16810 + }, + { + "epoch": 0.6494459245530716, + "grad_norm": 0.6847260594367981, + "learning_rate": 0.00015670617913175544, + "loss": 0.3681, + "step": 16820 + }, + { + "epoch": 0.6498320398470984, + "grad_norm": 3.0918118953704834, + "learning_rate": 0.00015668043811215362, + "loss": 0.3608, + "step": 16830 + }, + { + "epoch": 0.6502181551411251, + "grad_norm": 1.1295204162597656, + "learning_rate": 0.00015665469709255186, + "loss": 0.3189, + "step": 16840 + }, + { + "epoch": 0.6506042704351519, + "grad_norm": 1.803222417831421, + "learning_rate": 0.00015662895607295005, + "loss": 0.3138, + "step": 16850 + }, + { + "epoch": 0.6509903857291788, + "grad_norm": 3.157122850418091, + "learning_rate": 0.00015660321505334826, + "loss": 0.4342, + "step": 16860 + }, + { + "epoch": 0.6513765010232055, + "grad_norm": 2.6584184169769287, + "learning_rate": 0.00015657747403374648, + "loss": 0.211, + "step": 16870 + }, + { + "epoch": 0.6517626163172323, + "grad_norm": 0.763903021812439, + "learning_rate": 0.0001565517330141447, + "loss": 0.2873, + "step": 16880 + }, + { + "epoch": 0.6521487316112591, + "grad_norm": 2.5033602714538574, + "learning_rate": 0.00015652599199454293, + "loss": 0.4475, + "step": 16890 + }, + { + "epoch": 0.652534846905286, + "grad_norm": 2.088690757751465, + "learning_rate": 0.00015650025097494112, + "loss": 0.3022, + "step": 16900 + }, + { + "epoch": 0.6529209621993127, + "grad_norm": 1.596064567565918, + "learning_rate": 0.00015647450995533936, + "loss": 0.3771, + "step": 16910 + }, + { + "epoch": 0.6533070774933395, + "grad_norm": 1.2658660411834717, + "learning_rate": 0.00015644876893573754, + "loss": 0.4793, + "step": 16920 + }, + { + "epoch": 0.6536931927873663, + "grad_norm": 1.5343844890594482, + "learning_rate": 0.00015642302791613576, + "loss": 0.5026, + "step": 16930 + }, + { + "epoch": 0.6540793080813931, + "grad_norm": 0.4736674129962921, + "learning_rate": 0.000156397286896534, + "loss": 0.2269, + "step": 16940 + }, + { + "epoch": 0.6544654233754199, + "grad_norm": 0.05510171130299568, + "learning_rate": 0.00015637154587693218, + "loss": 0.2398, + "step": 16950 + }, + { + "epoch": 0.6548515386694467, + "grad_norm": 0.641941249370575, + "learning_rate": 0.00015634580485733042, + "loss": 0.3862, + "step": 16960 + }, + { + "epoch": 0.6552376539634734, + "grad_norm": 1.5418890714645386, + "learning_rate": 0.0001563200638377286, + "loss": 0.2688, + "step": 16970 + }, + { + "epoch": 0.6556237692575003, + "grad_norm": 3.46284818649292, + "learning_rate": 0.00015629432281812685, + "loss": 0.5199, + "step": 16980 + }, + { + "epoch": 0.6560098845515271, + "grad_norm": 0.3225530683994293, + "learning_rate": 0.00015626858179852504, + "loss": 0.5035, + "step": 16990 + }, + { + "epoch": 0.6563959998455539, + "grad_norm": 0.9385218620300293, + "learning_rate": 0.00015624284077892325, + "loss": 0.2618, + "step": 17000 + }, + { + "epoch": 0.6567821151395806, + "grad_norm": 0.8849124312400818, + "learning_rate": 0.0001562170997593215, + "loss": 0.3711, + "step": 17010 + }, + { + "epoch": 0.6571682304336075, + "grad_norm": 2.2706375122070312, + "learning_rate": 0.00015619135873971968, + "loss": 0.2666, + "step": 17020 + }, + { + "epoch": 0.6575543457276343, + "grad_norm": 2.1923744678497314, + "learning_rate": 0.00015616561772011792, + "loss": 0.2038, + "step": 17030 + }, + { + "epoch": 0.657940461021661, + "grad_norm": 1.2356051206588745, + "learning_rate": 0.0001561398767005161, + "loss": 0.4103, + "step": 17040 + }, + { + "epoch": 0.6583265763156879, + "grad_norm": 1.583095669746399, + "learning_rate": 0.00015611413568091434, + "loss": 0.4164, + "step": 17050 + }, + { + "epoch": 0.6587126916097147, + "grad_norm": 1.0823155641555786, + "learning_rate": 0.00015608839466131253, + "loss": 0.3898, + "step": 17060 + }, + { + "epoch": 0.6590988069037415, + "grad_norm": 1.9568531513214111, + "learning_rate": 0.00015606265364171074, + "loss": 0.3553, + "step": 17070 + }, + { + "epoch": 0.6594849221977682, + "grad_norm": 3.576362371444702, + "learning_rate": 0.00015603691262210898, + "loss": 0.2693, + "step": 17080 + }, + { + "epoch": 0.6598710374917951, + "grad_norm": 0.2432270646095276, + "learning_rate": 0.00015601117160250717, + "loss": 0.2993, + "step": 17090 + }, + { + "epoch": 0.6602571527858219, + "grad_norm": 1.5935213565826416, + "learning_rate": 0.0001559854305829054, + "loss": 0.3295, + "step": 17100 + }, + { + "epoch": 0.6606432680798486, + "grad_norm": 0.09780561178922653, + "learning_rate": 0.0001559596895633036, + "loss": 0.2891, + "step": 17110 + }, + { + "epoch": 0.6610293833738754, + "grad_norm": 0.5332283973693848, + "learning_rate": 0.00015593394854370184, + "loss": 0.372, + "step": 17120 + }, + { + "epoch": 0.6614154986679023, + "grad_norm": 1.1921123266220093, + "learning_rate": 0.00015590820752410005, + "loss": 0.3155, + "step": 17130 + }, + { + "epoch": 0.661801613961929, + "grad_norm": 0.35267120599746704, + "learning_rate": 0.00015588246650449823, + "loss": 0.3795, + "step": 17140 + }, + { + "epoch": 0.6621877292559558, + "grad_norm": 0.4876207411289215, + "learning_rate": 0.00015585672548489648, + "loss": 0.2717, + "step": 17150 + }, + { + "epoch": 0.6625738445499826, + "grad_norm": 0.9866208434104919, + "learning_rate": 0.00015583098446529466, + "loss": 0.4121, + "step": 17160 + }, + { + "epoch": 0.6629599598440095, + "grad_norm": 3.0264835357666016, + "learning_rate": 0.0001558052434456929, + "loss": 0.5356, + "step": 17170 + }, + { + "epoch": 0.6633460751380362, + "grad_norm": 2.4786953926086426, + "learning_rate": 0.0001557795024260911, + "loss": 0.2086, + "step": 17180 + }, + { + "epoch": 0.663732190432063, + "grad_norm": 2.3706555366516113, + "learning_rate": 0.00015575376140648933, + "loss": 0.5224, + "step": 17190 + }, + { + "epoch": 0.6641183057260898, + "grad_norm": 2.6375296115875244, + "learning_rate": 0.00015572802038688754, + "loss": 0.3625, + "step": 17200 + }, + { + "epoch": 0.6645044210201166, + "grad_norm": 0.5592703819274902, + "learning_rate": 0.00015570227936728573, + "loss": 0.3831, + "step": 17210 + }, + { + "epoch": 0.6648905363141434, + "grad_norm": 2.309683322906494, + "learning_rate": 0.00015567653834768397, + "loss": 0.4039, + "step": 17220 + }, + { + "epoch": 0.6652766516081702, + "grad_norm": 2.3134100437164307, + "learning_rate": 0.00015565079732808215, + "loss": 0.1904, + "step": 17230 + }, + { + "epoch": 0.665662766902197, + "grad_norm": 2.232910633087158, + "learning_rate": 0.0001556250563084804, + "loss": 0.34, + "step": 17240 + }, + { + "epoch": 0.6660488821962238, + "grad_norm": 0.4798373579978943, + "learning_rate": 0.0001555993152888786, + "loss": 0.3143, + "step": 17250 + }, + { + "epoch": 0.6664349974902506, + "grad_norm": 2.071753740310669, + "learning_rate": 0.00015557357426927682, + "loss": 0.2645, + "step": 17260 + }, + { + "epoch": 0.6668211127842774, + "grad_norm": 2.1930956840515137, + "learning_rate": 0.00015554783324967504, + "loss": 0.4144, + "step": 17270 + }, + { + "epoch": 0.6672072280783041, + "grad_norm": 1.7874137163162231, + "learning_rate": 0.00015552209223007325, + "loss": 0.273, + "step": 17280 + }, + { + "epoch": 0.667593343372331, + "grad_norm": 1.264596939086914, + "learning_rate": 0.00015549635121047146, + "loss": 0.4186, + "step": 17290 + }, + { + "epoch": 0.6679794586663578, + "grad_norm": 0.5612212419509888, + "learning_rate": 0.00015547061019086965, + "loss": 0.2802, + "step": 17300 + }, + { + "epoch": 0.6683655739603845, + "grad_norm": 1.3782585859298706, + "learning_rate": 0.0001554448691712679, + "loss": 0.3712, + "step": 17310 + }, + { + "epoch": 0.6687516892544114, + "grad_norm": 1.5178605318069458, + "learning_rate": 0.0001554191281516661, + "loss": 0.1694, + "step": 17320 + }, + { + "epoch": 0.6691378045484382, + "grad_norm": 2.1221604347229004, + "learning_rate": 0.00015539338713206432, + "loss": 0.4418, + "step": 17330 + }, + { + "epoch": 0.669523919842465, + "grad_norm": 1.570734977722168, + "learning_rate": 0.00015536764611246253, + "loss": 0.4037, + "step": 17340 + }, + { + "epoch": 0.6699100351364917, + "grad_norm": 0.6928157806396484, + "learning_rate": 0.00015534190509286074, + "loss": 0.5293, + "step": 17350 + }, + { + "epoch": 0.6702961504305186, + "grad_norm": 0.8526401519775391, + "learning_rate": 0.00015531616407325895, + "loss": 0.348, + "step": 17360 + }, + { + "epoch": 0.6706822657245454, + "grad_norm": 1.7482202053070068, + "learning_rate": 0.00015529042305365714, + "loss": 0.352, + "step": 17370 + }, + { + "epoch": 0.6710683810185721, + "grad_norm": 1.724870204925537, + "learning_rate": 0.00015526468203405538, + "loss": 0.3589, + "step": 17380 + }, + { + "epoch": 0.6714544963125989, + "grad_norm": 3.125180721282959, + "learning_rate": 0.0001552389410144536, + "loss": 0.3063, + "step": 17390 + }, + { + "epoch": 0.6718406116066258, + "grad_norm": 2.0817360877990723, + "learning_rate": 0.0001552131999948518, + "loss": 0.2217, + "step": 17400 + }, + { + "epoch": 0.6722267269006525, + "grad_norm": 0.12367000430822372, + "learning_rate": 0.00015518745897525002, + "loss": 0.1691, + "step": 17410 + }, + { + "epoch": 0.6726128421946793, + "grad_norm": 0.23093344271183014, + "learning_rate": 0.00015516171795564823, + "loss": 0.2496, + "step": 17420 + }, + { + "epoch": 0.6729989574887061, + "grad_norm": 3.1588997840881348, + "learning_rate": 0.00015513597693604645, + "loss": 0.2868, + "step": 17430 + }, + { + "epoch": 0.673385072782733, + "grad_norm": 1.471999168395996, + "learning_rate": 0.00015511023591644466, + "loss": 0.2785, + "step": 17440 + }, + { + "epoch": 0.6737711880767597, + "grad_norm": 0.4500691294670105, + "learning_rate": 0.00015508449489684287, + "loss": 0.3218, + "step": 17450 + }, + { + "epoch": 0.6741573033707865, + "grad_norm": 2.65533709526062, + "learning_rate": 0.0001550587538772411, + "loss": 0.3194, + "step": 17460 + }, + { + "epoch": 0.6745434186648133, + "grad_norm": 0.45398348569869995, + "learning_rate": 0.0001550330128576393, + "loss": 0.199, + "step": 17470 + }, + { + "epoch": 0.6749295339588401, + "grad_norm": 0.21518200635910034, + "learning_rate": 0.00015500727183803751, + "loss": 0.3043, + "step": 17480 + }, + { + "epoch": 0.6753156492528669, + "grad_norm": 0.13117246329784393, + "learning_rate": 0.00015498153081843573, + "loss": 0.1872, + "step": 17490 + }, + { + "epoch": 0.6757017645468937, + "grad_norm": 0.4857695996761322, + "learning_rate": 0.00015495578979883394, + "loss": 0.5992, + "step": 17500 + }, + { + "epoch": 0.6760878798409204, + "grad_norm": 2.4992752075195312, + "learning_rate": 0.00015493004877923215, + "loss": 0.5057, + "step": 17510 + }, + { + "epoch": 0.6764739951349473, + "grad_norm": 1.9614732265472412, + "learning_rate": 0.00015490430775963037, + "loss": 0.3169, + "step": 17520 + }, + { + "epoch": 0.6768601104289741, + "grad_norm": 0.14168275892734528, + "learning_rate": 0.00015487856674002858, + "loss": 0.271, + "step": 17530 + }, + { + "epoch": 0.6772462257230009, + "grad_norm": 4.064804553985596, + "learning_rate": 0.0001548528257204268, + "loss": 0.3316, + "step": 17540 + }, + { + "epoch": 0.6776323410170277, + "grad_norm": 3.2959964275360107, + "learning_rate": 0.000154827084700825, + "loss": 0.5148, + "step": 17550 + }, + { + "epoch": 0.6780184563110545, + "grad_norm": 0.6234021186828613, + "learning_rate": 0.00015480134368122322, + "loss": 0.362, + "step": 17560 + }, + { + "epoch": 0.6784045716050813, + "grad_norm": 0.642573356628418, + "learning_rate": 0.00015477560266162143, + "loss": 0.2982, + "step": 17570 + }, + { + "epoch": 0.678790686899108, + "grad_norm": 1.098667025566101, + "learning_rate": 0.00015474986164201965, + "loss": 0.1875, + "step": 17580 + }, + { + "epoch": 0.6791768021931349, + "grad_norm": 2.378192186355591, + "learning_rate": 0.00015472412062241786, + "loss": 0.2533, + "step": 17590 + }, + { + "epoch": 0.6795629174871617, + "grad_norm": 1.1783161163330078, + "learning_rate": 0.00015469837960281607, + "loss": 0.402, + "step": 17600 + }, + { + "epoch": 0.6799490327811885, + "grad_norm": 0.1861846148967743, + "learning_rate": 0.0001546726385832143, + "loss": 0.2481, + "step": 17610 + }, + { + "epoch": 0.6803351480752152, + "grad_norm": 0.5785403847694397, + "learning_rate": 0.0001546468975636125, + "loss": 0.375, + "step": 17620 + }, + { + "epoch": 0.6807212633692421, + "grad_norm": 1.9201544523239136, + "learning_rate": 0.00015462115654401071, + "loss": 0.2757, + "step": 17630 + }, + { + "epoch": 0.6811073786632689, + "grad_norm": 2.461735963821411, + "learning_rate": 0.00015459541552440893, + "loss": 0.2193, + "step": 17640 + }, + { + "epoch": 0.6814934939572956, + "grad_norm": 2.007638454437256, + "learning_rate": 0.00015456967450480714, + "loss": 0.3561, + "step": 17650 + }, + { + "epoch": 0.6818796092513224, + "grad_norm": 1.3581938743591309, + "learning_rate": 0.00015454393348520535, + "loss": 0.3736, + "step": 17660 + }, + { + "epoch": 0.6822657245453493, + "grad_norm": 0.5637246966362, + "learning_rate": 0.00015451819246560357, + "loss": 0.3116, + "step": 17670 + }, + { + "epoch": 0.682651839839376, + "grad_norm": 1.409740924835205, + "learning_rate": 0.00015449245144600178, + "loss": 0.3281, + "step": 17680 + }, + { + "epoch": 0.6830379551334028, + "grad_norm": 3.2064149379730225, + "learning_rate": 0.0001544667104264, + "loss": 0.4427, + "step": 17690 + }, + { + "epoch": 0.6834240704274296, + "grad_norm": 0.9369992613792419, + "learning_rate": 0.0001544409694067982, + "loss": 0.3424, + "step": 17700 + }, + { + "epoch": 0.6838101857214565, + "grad_norm": 2.4149889945983887, + "learning_rate": 0.00015441522838719642, + "loss": 0.4074, + "step": 17710 + }, + { + "epoch": 0.6841963010154832, + "grad_norm": 0.688360869884491, + "learning_rate": 0.00015438948736759463, + "loss": 0.2205, + "step": 17720 + }, + { + "epoch": 0.68458241630951, + "grad_norm": 2.1444098949432373, + "learning_rate": 0.00015436374634799285, + "loss": 0.5942, + "step": 17730 + }, + { + "epoch": 0.6849685316035368, + "grad_norm": 1.8053444623947144, + "learning_rate": 0.00015433800532839106, + "loss": 0.1665, + "step": 17740 + }, + { + "epoch": 0.6853546468975636, + "grad_norm": 3.5637879371643066, + "learning_rate": 0.0001543122643087893, + "loss": 0.4972, + "step": 17750 + }, + { + "epoch": 0.6857407621915904, + "grad_norm": 1.846845030784607, + "learning_rate": 0.0001542865232891875, + "loss": 0.4105, + "step": 17760 + }, + { + "epoch": 0.6861268774856172, + "grad_norm": 2.2459189891815186, + "learning_rate": 0.0001542607822695857, + "loss": 0.253, + "step": 17770 + }, + { + "epoch": 0.686512992779644, + "grad_norm": 2.3160414695739746, + "learning_rate": 0.00015423504124998391, + "loss": 0.1905, + "step": 17780 + }, + { + "epoch": 0.6868991080736708, + "grad_norm": 1.2804152965545654, + "learning_rate": 0.00015420930023038213, + "loss": 0.2283, + "step": 17790 + }, + { + "epoch": 0.6872852233676976, + "grad_norm": 1.7174758911132812, + "learning_rate": 0.00015418355921078034, + "loss": 0.49, + "step": 17800 + }, + { + "epoch": 0.6876713386617244, + "grad_norm": 3.057098627090454, + "learning_rate": 0.00015415781819117855, + "loss": 0.4398, + "step": 17810 + }, + { + "epoch": 0.6880574539557512, + "grad_norm": 0.9112808704376221, + "learning_rate": 0.0001541320771715768, + "loss": 0.203, + "step": 17820 + }, + { + "epoch": 0.688443569249778, + "grad_norm": 2.899599313735962, + "learning_rate": 0.00015410633615197498, + "loss": 0.4212, + "step": 17830 + }, + { + "epoch": 0.6888296845438048, + "grad_norm": 1.8084157705307007, + "learning_rate": 0.0001540805951323732, + "loss": 0.4948, + "step": 17840 + }, + { + "epoch": 0.6892157998378315, + "grad_norm": 1.2151083946228027, + "learning_rate": 0.0001540548541127714, + "loss": 0.4254, + "step": 17850 + }, + { + "epoch": 0.6896019151318584, + "grad_norm": 1.20271897315979, + "learning_rate": 0.00015402911309316962, + "loss": 0.1753, + "step": 17860 + }, + { + "epoch": 0.6899880304258852, + "grad_norm": 0.7688419222831726, + "learning_rate": 0.00015400337207356783, + "loss": 0.4066, + "step": 17870 + }, + { + "epoch": 0.690374145719912, + "grad_norm": 0.8648087978363037, + "learning_rate": 0.00015397763105396605, + "loss": 0.5405, + "step": 17880 + }, + { + "epoch": 0.6907602610139387, + "grad_norm": 1.5501036643981934, + "learning_rate": 0.0001539518900343643, + "loss": 0.3761, + "step": 17890 + }, + { + "epoch": 0.6911463763079656, + "grad_norm": 0.5476267337799072, + "learning_rate": 0.00015392614901476247, + "loss": 0.2256, + "step": 17900 + }, + { + "epoch": 0.6915324916019924, + "grad_norm": 2.0248584747314453, + "learning_rate": 0.0001539004079951607, + "loss": 0.5449, + "step": 17910 + }, + { + "epoch": 0.6919186068960191, + "grad_norm": 1.688596248626709, + "learning_rate": 0.0001538746669755589, + "loss": 0.262, + "step": 17920 + }, + { + "epoch": 0.6923047221900459, + "grad_norm": 0.42646175622940063, + "learning_rate": 0.00015384892595595711, + "loss": 0.3714, + "step": 17930 + }, + { + "epoch": 0.6926908374840728, + "grad_norm": 0.9620506167411804, + "learning_rate": 0.00015382318493635535, + "loss": 0.272, + "step": 17940 + }, + { + "epoch": 0.6930769527780996, + "grad_norm": 1.6859287023544312, + "learning_rate": 0.00015379744391675354, + "loss": 0.3123, + "step": 17950 + }, + { + "epoch": 0.6934630680721263, + "grad_norm": 0.6281775832176208, + "learning_rate": 0.00015377170289715178, + "loss": 0.2808, + "step": 17960 + }, + { + "epoch": 0.6938491833661531, + "grad_norm": 3.756242036819458, + "learning_rate": 0.00015374596187754997, + "loss": 0.4971, + "step": 17970 + }, + { + "epoch": 0.69423529866018, + "grad_norm": 0.6022955775260925, + "learning_rate": 0.0001537202208579482, + "loss": 0.2918, + "step": 17980 + }, + { + "epoch": 0.6946214139542067, + "grad_norm": 0.6843704581260681, + "learning_rate": 0.0001536944798383464, + "loss": 0.476, + "step": 17990 + }, + { + "epoch": 0.6950075292482335, + "grad_norm": 1.0234850645065308, + "learning_rate": 0.0001536687388187446, + "loss": 0.2101, + "step": 18000 + }, + { + "epoch": 0.6953936445422603, + "grad_norm": 1.0228936672210693, + "learning_rate": 0.00015364299779914285, + "loss": 0.3958, + "step": 18010 + }, + { + "epoch": 0.6957797598362871, + "grad_norm": 1.1152328252792358, + "learning_rate": 0.00015361725677954103, + "loss": 0.2967, + "step": 18020 + }, + { + "epoch": 0.6961658751303139, + "grad_norm": 1.7190260887145996, + "learning_rate": 0.00015359151575993927, + "loss": 0.5281, + "step": 18030 + }, + { + "epoch": 0.6965519904243407, + "grad_norm": 0.6654171943664551, + "learning_rate": 0.00015356577474033746, + "loss": 0.3467, + "step": 18040 + }, + { + "epoch": 0.6969381057183675, + "grad_norm": 0.3305549621582031, + "learning_rate": 0.0001535400337207357, + "loss": 0.3576, + "step": 18050 + }, + { + "epoch": 0.6973242210123943, + "grad_norm": 0.3116997480392456, + "learning_rate": 0.0001535142927011339, + "loss": 0.5372, + "step": 18060 + }, + { + "epoch": 0.6977103363064211, + "grad_norm": 0.4224954843521118, + "learning_rate": 0.0001534885516815321, + "loss": 0.3866, + "step": 18070 + }, + { + "epoch": 0.6980964516004479, + "grad_norm": 4.249162197113037, + "learning_rate": 0.00015346281066193034, + "loss": 0.4236, + "step": 18080 + }, + { + "epoch": 0.6984825668944747, + "grad_norm": 1.109113335609436, + "learning_rate": 0.00015343706964232853, + "loss": 0.2882, + "step": 18090 + }, + { + "epoch": 0.6988686821885015, + "grad_norm": 1.3546028137207031, + "learning_rate": 0.00015341132862272677, + "loss": 0.3432, + "step": 18100 + }, + { + "epoch": 0.6992547974825283, + "grad_norm": 2.943016290664673, + "learning_rate": 0.00015338558760312495, + "loss": 0.3287, + "step": 18110 + }, + { + "epoch": 0.699640912776555, + "grad_norm": 1.1259021759033203, + "learning_rate": 0.0001533598465835232, + "loss": 0.4613, + "step": 18120 + }, + { + "epoch": 0.7000270280705819, + "grad_norm": 1.4867910146713257, + "learning_rate": 0.0001533341055639214, + "loss": 0.4309, + "step": 18130 + }, + { + "epoch": 0.7004131433646087, + "grad_norm": 2.6913414001464844, + "learning_rate": 0.0001533083645443196, + "loss": 0.2154, + "step": 18140 + }, + { + "epoch": 0.7007992586586355, + "grad_norm": 1.495466947555542, + "learning_rate": 0.00015328262352471783, + "loss": 0.3207, + "step": 18150 + }, + { + "epoch": 0.7011853739526622, + "grad_norm": 1.023193120956421, + "learning_rate": 0.00015325688250511602, + "loss": 0.2067, + "step": 18160 + }, + { + "epoch": 0.7015714892466891, + "grad_norm": 1.603235125541687, + "learning_rate": 0.00015323114148551426, + "loss": 0.4577, + "step": 18170 + }, + { + "epoch": 0.7019576045407159, + "grad_norm": 0.5976241230964661, + "learning_rate": 0.00015320540046591245, + "loss": 0.2282, + "step": 18180 + }, + { + "epoch": 0.7023437198347426, + "grad_norm": 2.561659574508667, + "learning_rate": 0.0001531796594463107, + "loss": 0.4045, + "step": 18190 + }, + { + "epoch": 0.7027298351287694, + "grad_norm": 1.3893495798110962, + "learning_rate": 0.0001531539184267089, + "loss": 0.2419, + "step": 18200 + }, + { + "epoch": 0.7031159504227963, + "grad_norm": 0.7786352038383484, + "learning_rate": 0.00015312817740710709, + "loss": 0.1653, + "step": 18210 + }, + { + "epoch": 0.703502065716823, + "grad_norm": 0.6525956988334656, + "learning_rate": 0.00015310243638750533, + "loss": 0.5418, + "step": 18220 + }, + { + "epoch": 0.7038881810108498, + "grad_norm": 0.38933584094047546, + "learning_rate": 0.0001530766953679035, + "loss": 0.2952, + "step": 18230 + }, + { + "epoch": 0.7042742963048766, + "grad_norm": 2.0752692222595215, + "learning_rate": 0.00015305095434830175, + "loss": 0.211, + "step": 18240 + }, + { + "epoch": 0.7046604115989035, + "grad_norm": 0.9095730781555176, + "learning_rate": 0.00015302521332869997, + "loss": 0.2723, + "step": 18250 + }, + { + "epoch": 0.7050465268929302, + "grad_norm": 1.6840119361877441, + "learning_rate": 0.00015299947230909818, + "loss": 0.3362, + "step": 18260 + }, + { + "epoch": 0.705432642186957, + "grad_norm": 2.0353269577026367, + "learning_rate": 0.0001529737312894964, + "loss": 0.2407, + "step": 18270 + }, + { + "epoch": 0.7058187574809838, + "grad_norm": 3.0865590572357178, + "learning_rate": 0.00015294799026989458, + "loss": 0.3426, + "step": 18280 + }, + { + "epoch": 0.7062048727750107, + "grad_norm": 1.6488090753555298, + "learning_rate": 0.00015292224925029282, + "loss": 0.4275, + "step": 18290 + }, + { + "epoch": 0.7065909880690374, + "grad_norm": 0.5494143962860107, + "learning_rate": 0.000152896508230691, + "loss": 0.412, + "step": 18300 + }, + { + "epoch": 0.7069771033630642, + "grad_norm": 3.111301898956299, + "learning_rate": 0.00015287076721108925, + "loss": 0.4615, + "step": 18310 + }, + { + "epoch": 0.707363218657091, + "grad_norm": 1.74229097366333, + "learning_rate": 0.00015284502619148746, + "loss": 0.3194, + "step": 18320 + }, + { + "epoch": 0.7077493339511178, + "grad_norm": 1.8455474376678467, + "learning_rate": 0.00015281928517188567, + "loss": 0.2817, + "step": 18330 + }, + { + "epoch": 0.7081354492451446, + "grad_norm": 1.778723120689392, + "learning_rate": 0.0001527935441522839, + "loss": 0.441, + "step": 18340 + }, + { + "epoch": 0.7085215645391714, + "grad_norm": 0.7885593771934509, + "learning_rate": 0.00015276780313268207, + "loss": 0.3121, + "step": 18350 + }, + { + "epoch": 0.7089076798331982, + "grad_norm": 1.9262609481811523, + "learning_rate": 0.0001527420621130803, + "loss": 0.3271, + "step": 18360 + }, + { + "epoch": 0.709293795127225, + "grad_norm": 0.10278096795082092, + "learning_rate": 0.0001527163210934785, + "loss": 0.2602, + "step": 18370 + }, + { + "epoch": 0.7096799104212518, + "grad_norm": 1.2394765615463257, + "learning_rate": 0.00015269058007387674, + "loss": 0.2641, + "step": 18380 + }, + { + "epoch": 0.7100660257152785, + "grad_norm": 2.0335285663604736, + "learning_rate": 0.00015266483905427495, + "loss": 0.1926, + "step": 18390 + }, + { + "epoch": 0.7104521410093054, + "grad_norm": 7.205105781555176, + "learning_rate": 0.00015263909803467317, + "loss": 0.4772, + "step": 18400 + }, + { + "epoch": 0.7108382563033322, + "grad_norm": 1.1549599170684814, + "learning_rate": 0.00015261335701507138, + "loss": 0.2775, + "step": 18410 + }, + { + "epoch": 0.711224371597359, + "grad_norm": 2.8363780975341797, + "learning_rate": 0.00015258761599546957, + "loss": 0.3533, + "step": 18420 + }, + { + "epoch": 0.7116104868913857, + "grad_norm": 0.2606666684150696, + "learning_rate": 0.0001525618749758678, + "loss": 0.1421, + "step": 18430 + }, + { + "epoch": 0.7119966021854126, + "grad_norm": 1.2934225797653198, + "learning_rate": 0.00015253613395626602, + "loss": 0.3066, + "step": 18440 + }, + { + "epoch": 0.7123827174794394, + "grad_norm": 3.8246026039123535, + "learning_rate": 0.00015251039293666423, + "loss": 0.534, + "step": 18450 + }, + { + "epoch": 0.7127688327734661, + "grad_norm": 2.2535433769226074, + "learning_rate": 0.00015248465191706245, + "loss": 0.4795, + "step": 18460 + }, + { + "epoch": 0.7131549480674929, + "grad_norm": 1.6749187707901, + "learning_rate": 0.00015245891089746066, + "loss": 0.4091, + "step": 18470 + }, + { + "epoch": 0.7135410633615198, + "grad_norm": 1.1795039176940918, + "learning_rate": 0.00015243316987785887, + "loss": 0.2612, + "step": 18480 + }, + { + "epoch": 0.7139271786555466, + "grad_norm": 1.9280221462249756, + "learning_rate": 0.00015240742885825709, + "loss": 0.2068, + "step": 18490 + }, + { + "epoch": 0.7143132939495733, + "grad_norm": 1.1188548803329468, + "learning_rate": 0.0001523816878386553, + "loss": 0.3193, + "step": 18500 + }, + { + "epoch": 0.7146994092436001, + "grad_norm": 0.2429720014333725, + "learning_rate": 0.0001523559468190535, + "loss": 0.3129, + "step": 18510 + }, + { + "epoch": 0.715085524537627, + "grad_norm": 4.09410285949707, + "learning_rate": 0.00015233020579945173, + "loss": 0.2437, + "step": 18520 + }, + { + "epoch": 0.7154716398316537, + "grad_norm": 2.3252813816070557, + "learning_rate": 0.00015230446477984994, + "loss": 0.4649, + "step": 18530 + }, + { + "epoch": 0.7158577551256805, + "grad_norm": 0.5725727677345276, + "learning_rate": 0.00015227872376024815, + "loss": 0.3291, + "step": 18540 + }, + { + "epoch": 0.7162438704197073, + "grad_norm": 0.9253637194633484, + "learning_rate": 0.00015225298274064637, + "loss": 0.3486, + "step": 18550 + }, + { + "epoch": 0.7166299857137342, + "grad_norm": 2.3353309631347656, + "learning_rate": 0.00015222724172104458, + "loss": 0.253, + "step": 18560 + }, + { + "epoch": 0.7170161010077609, + "grad_norm": 0.7312389016151428, + "learning_rate": 0.0001522015007014428, + "loss": 0.2817, + "step": 18570 + }, + { + "epoch": 0.7174022163017877, + "grad_norm": 0.6564128994941711, + "learning_rate": 0.000152175759681841, + "loss": 0.2896, + "step": 18580 + }, + { + "epoch": 0.7177883315958145, + "grad_norm": 3.4619979858398438, + "learning_rate": 0.00015215001866223922, + "loss": 0.5028, + "step": 18590 + }, + { + "epoch": 0.7181744468898413, + "grad_norm": 6.910060882568359, + "learning_rate": 0.00015212427764263743, + "loss": 0.2467, + "step": 18600 + }, + { + "epoch": 0.7185605621838681, + "grad_norm": 2.022186279296875, + "learning_rate": 0.00015209853662303565, + "loss": 0.3406, + "step": 18610 + }, + { + "epoch": 0.7189466774778949, + "grad_norm": 1.2240760326385498, + "learning_rate": 0.00015207279560343386, + "loss": 0.3391, + "step": 18620 + }, + { + "epoch": 0.7193327927719217, + "grad_norm": 0.7356148958206177, + "learning_rate": 0.00015204705458383207, + "loss": 0.2679, + "step": 18630 + }, + { + "epoch": 0.7197189080659485, + "grad_norm": 0.963387131690979, + "learning_rate": 0.00015202131356423029, + "loss": 0.4088, + "step": 18640 + }, + { + "epoch": 0.7201050233599753, + "grad_norm": 3.0437800884246826, + "learning_rate": 0.0001519955725446285, + "loss": 0.2591, + "step": 18650 + }, + { + "epoch": 0.720491138654002, + "grad_norm": 2.5874569416046143, + "learning_rate": 0.0001519698315250267, + "loss": 0.2377, + "step": 18660 + }, + { + "epoch": 0.7208772539480289, + "grad_norm": 2.3215808868408203, + "learning_rate": 0.00015194409050542493, + "loss": 0.5335, + "step": 18670 + }, + { + "epoch": 0.7212633692420557, + "grad_norm": 1.9501638412475586, + "learning_rate": 0.00015191834948582314, + "loss": 0.4657, + "step": 18680 + }, + { + "epoch": 0.7216494845360825, + "grad_norm": 1.8396021127700806, + "learning_rate": 0.00015189260846622135, + "loss": 0.2719, + "step": 18690 + }, + { + "epoch": 0.7220355998301092, + "grad_norm": 0.9337745904922485, + "learning_rate": 0.00015186686744661957, + "loss": 0.3672, + "step": 18700 + }, + { + "epoch": 0.7224217151241361, + "grad_norm": 1.892098069190979, + "learning_rate": 0.00015184112642701778, + "loss": 0.6085, + "step": 18710 + }, + { + "epoch": 0.7228078304181629, + "grad_norm": 1.051630973815918, + "learning_rate": 0.000151815385407416, + "loss": 0.2422, + "step": 18720 + }, + { + "epoch": 0.7231939457121896, + "grad_norm": 0.8714147210121155, + "learning_rate": 0.0001517896443878142, + "loss": 0.4046, + "step": 18730 + }, + { + "epoch": 0.7235800610062164, + "grad_norm": 0.5002617835998535, + "learning_rate": 0.00015176390336821242, + "loss": 0.3708, + "step": 18740 + }, + { + "epoch": 0.7239661763002433, + "grad_norm": 1.3960262537002563, + "learning_rate": 0.00015173816234861066, + "loss": 0.3206, + "step": 18750 + }, + { + "epoch": 0.7243522915942701, + "grad_norm": 0.7899012565612793, + "learning_rate": 0.00015171242132900885, + "loss": 0.2875, + "step": 18760 + }, + { + "epoch": 0.7247384068882968, + "grad_norm": 0.7216291427612305, + "learning_rate": 0.00015168668030940706, + "loss": 0.3885, + "step": 18770 + }, + { + "epoch": 0.7251245221823236, + "grad_norm": 0.36028966307640076, + "learning_rate": 0.00015166093928980527, + "loss": 0.3542, + "step": 18780 + }, + { + "epoch": 0.7255106374763505, + "grad_norm": 1.378724455833435, + "learning_rate": 0.00015163519827020349, + "loss": 0.2149, + "step": 18790 + }, + { + "epoch": 0.7258967527703772, + "grad_norm": 1.544819712638855, + "learning_rate": 0.0001516094572506017, + "loss": 0.278, + "step": 18800 + }, + { + "epoch": 0.726282868064404, + "grad_norm": 0.5819025039672852, + "learning_rate": 0.0001515837162309999, + "loss": 0.2161, + "step": 18810 + }, + { + "epoch": 0.7266689833584308, + "grad_norm": 0.8350955843925476, + "learning_rate": 0.00015155797521139815, + "loss": 0.4134, + "step": 18820 + }, + { + "epoch": 0.7270550986524577, + "grad_norm": 2.4110195636749268, + "learning_rate": 0.00015153223419179634, + "loss": 0.3928, + "step": 18830 + }, + { + "epoch": 0.7274412139464844, + "grad_norm": 0.7913835048675537, + "learning_rate": 0.00015150649317219455, + "loss": 0.2336, + "step": 18840 + }, + { + "epoch": 0.7278273292405112, + "grad_norm": 2.13431978225708, + "learning_rate": 0.00015148075215259277, + "loss": 0.4758, + "step": 18850 + }, + { + "epoch": 0.7282134445345381, + "grad_norm": 1.2756295204162598, + "learning_rate": 0.00015145501113299098, + "loss": 0.3139, + "step": 18860 + }, + { + "epoch": 0.7285995598285648, + "grad_norm": 0.3661370277404785, + "learning_rate": 0.0001514292701133892, + "loss": 0.3913, + "step": 18870 + }, + { + "epoch": 0.7289856751225916, + "grad_norm": 1.2238267660140991, + "learning_rate": 0.0001514035290937874, + "loss": 0.1893, + "step": 18880 + }, + { + "epoch": 0.7293717904166184, + "grad_norm": 2.8876595497131348, + "learning_rate": 0.00015137778807418565, + "loss": 0.2567, + "step": 18890 + }, + { + "epoch": 0.7297579057106452, + "grad_norm": 1.248967170715332, + "learning_rate": 0.00015135204705458383, + "loss": 0.2004, + "step": 18900 + }, + { + "epoch": 0.730144021004672, + "grad_norm": 0.9446873068809509, + "learning_rate": 0.00015132630603498205, + "loss": 0.386, + "step": 18910 + }, + { + "epoch": 0.7305301362986988, + "grad_norm": 4.592974662780762, + "learning_rate": 0.00015130056501538026, + "loss": 0.4549, + "step": 18920 + }, + { + "epoch": 0.7309162515927256, + "grad_norm": 0.8000105619430542, + "learning_rate": 0.00015127482399577847, + "loss": 0.2795, + "step": 18930 + }, + { + "epoch": 0.7313023668867524, + "grad_norm": 0.5600059628486633, + "learning_rate": 0.0001512490829761767, + "loss": 0.3682, + "step": 18940 + }, + { + "epoch": 0.7316884821807792, + "grad_norm": 0.4072086811065674, + "learning_rate": 0.0001512233419565749, + "loss": 0.3186, + "step": 18950 + }, + { + "epoch": 0.732074597474806, + "grad_norm": 3.387422561645508, + "learning_rate": 0.00015119760093697314, + "loss": 0.3195, + "step": 18960 + }, + { + "epoch": 0.7324607127688327, + "grad_norm": 0.5224191546440125, + "learning_rate": 0.00015117185991737133, + "loss": 0.1792, + "step": 18970 + }, + { + "epoch": 0.7328468280628596, + "grad_norm": 0.5431543588638306, + "learning_rate": 0.00015114611889776954, + "loss": 0.2775, + "step": 18980 + }, + { + "epoch": 0.7332329433568864, + "grad_norm": 0.13088488578796387, + "learning_rate": 0.00015112037787816775, + "loss": 0.3511, + "step": 18990 + }, + { + "epoch": 0.7336190586509131, + "grad_norm": 0.6414417624473572, + "learning_rate": 0.00015109463685856597, + "loss": 0.4042, + "step": 19000 + }, + { + "epoch": 0.7340051739449399, + "grad_norm": 0.2358855903148651, + "learning_rate": 0.0001510688958389642, + "loss": 0.2567, + "step": 19010 + }, + { + "epoch": 0.7343912892389668, + "grad_norm": 1.343703031539917, + "learning_rate": 0.0001510431548193624, + "loss": 0.3658, + "step": 19020 + }, + { + "epoch": 0.7347774045329936, + "grad_norm": 2.5982301235198975, + "learning_rate": 0.00015101741379976063, + "loss": 0.3664, + "step": 19030 + }, + { + "epoch": 0.7351635198270203, + "grad_norm": 1.3639850616455078, + "learning_rate": 0.00015099167278015882, + "loss": 0.1809, + "step": 19040 + }, + { + "epoch": 0.7355496351210471, + "grad_norm": 1.322572946548462, + "learning_rate": 0.00015096593176055703, + "loss": 0.213, + "step": 19050 + }, + { + "epoch": 0.735935750415074, + "grad_norm": 0.6858059763908386, + "learning_rate": 0.00015094019074095527, + "loss": 0.1494, + "step": 19060 + }, + { + "epoch": 0.7363218657091007, + "grad_norm": 3.9815866947174072, + "learning_rate": 0.00015091444972135346, + "loss": 0.4879, + "step": 19070 + }, + { + "epoch": 0.7367079810031275, + "grad_norm": 0.5155348777770996, + "learning_rate": 0.0001508887087017517, + "loss": 0.1951, + "step": 19080 + }, + { + "epoch": 0.7370940962971543, + "grad_norm": 1.1120082139968872, + "learning_rate": 0.00015086296768214988, + "loss": 0.3156, + "step": 19090 + }, + { + "epoch": 0.7374802115911812, + "grad_norm": 2.1396732330322266, + "learning_rate": 0.00015083722666254813, + "loss": 0.3528, + "step": 19100 + }, + { + "epoch": 0.7378663268852079, + "grad_norm": 1.8543074131011963, + "learning_rate": 0.0001508114856429463, + "loss": 0.2778, + "step": 19110 + }, + { + "epoch": 0.7382524421792347, + "grad_norm": 0.2955397963523865, + "learning_rate": 0.00015078574462334452, + "loss": 0.3267, + "step": 19120 + }, + { + "epoch": 0.7386385574732616, + "grad_norm": 0.6277685165405273, + "learning_rate": 0.00015076000360374277, + "loss": 0.3316, + "step": 19130 + }, + { + "epoch": 0.7390246727672883, + "grad_norm": 1.4339113235473633, + "learning_rate": 0.00015073426258414095, + "loss": 0.3227, + "step": 19140 + }, + { + "epoch": 0.7394107880613151, + "grad_norm": 2.74206805229187, + "learning_rate": 0.0001507085215645392, + "loss": 0.2086, + "step": 19150 + }, + { + "epoch": 0.7397969033553419, + "grad_norm": 1.0108954906463623, + "learning_rate": 0.00015068278054493738, + "loss": 0.2697, + "step": 19160 + }, + { + "epoch": 0.7401830186493688, + "grad_norm": 1.726650357246399, + "learning_rate": 0.00015065703952533562, + "loss": 0.1312, + "step": 19170 + }, + { + "epoch": 0.7405691339433955, + "grad_norm": 1.8226735591888428, + "learning_rate": 0.0001506312985057338, + "loss": 0.4488, + "step": 19180 + }, + { + "epoch": 0.7409552492374223, + "grad_norm": 1.4812517166137695, + "learning_rate": 0.00015060555748613205, + "loss": 0.472, + "step": 19190 + }, + { + "epoch": 0.741341364531449, + "grad_norm": 1.184036374092102, + "learning_rate": 0.00015057981646653026, + "loss": 0.1342, + "step": 19200 + }, + { + "epoch": 0.7417274798254759, + "grad_norm": 2.925368309020996, + "learning_rate": 0.00015055407544692844, + "loss": 0.4114, + "step": 19210 + }, + { + "epoch": 0.7421135951195027, + "grad_norm": 1.4891862869262695, + "learning_rate": 0.00015052833442732669, + "loss": 0.3454, + "step": 19220 + }, + { + "epoch": 0.7424997104135295, + "grad_norm": 2.9221529960632324, + "learning_rate": 0.00015050259340772487, + "loss": 0.4622, + "step": 19230 + }, + { + "epoch": 0.7428858257075562, + "grad_norm": 1.3214635848999023, + "learning_rate": 0.0001504768523881231, + "loss": 0.2783, + "step": 19240 + }, + { + "epoch": 0.7432719410015831, + "grad_norm": 1.1919734477996826, + "learning_rate": 0.00015045111136852133, + "loss": 0.3667, + "step": 19250 + }, + { + "epoch": 0.7436580562956099, + "grad_norm": 1.7075424194335938, + "learning_rate": 0.00015042537034891954, + "loss": 0.1867, + "step": 19260 + }, + { + "epoch": 0.7440441715896366, + "grad_norm": 0.6810876727104187, + "learning_rate": 0.00015039962932931775, + "loss": 0.2231, + "step": 19270 + }, + { + "epoch": 0.7444302868836634, + "grad_norm": 1.3421598672866821, + "learning_rate": 0.00015037388830971594, + "loss": 0.2138, + "step": 19280 + }, + { + "epoch": 0.7448164021776903, + "grad_norm": 1.5983080863952637, + "learning_rate": 0.00015034814729011418, + "loss": 0.2799, + "step": 19290 + }, + { + "epoch": 0.7452025174717171, + "grad_norm": 2.0319182872772217, + "learning_rate": 0.00015032240627051236, + "loss": 0.213, + "step": 19300 + }, + { + "epoch": 0.7455886327657438, + "grad_norm": 0.6433222889900208, + "learning_rate": 0.0001502966652509106, + "loss": 0.3604, + "step": 19310 + }, + { + "epoch": 0.7459747480597706, + "grad_norm": 1.1373825073242188, + "learning_rate": 0.00015027092423130882, + "loss": 0.3007, + "step": 19320 + }, + { + "epoch": 0.7463608633537975, + "grad_norm": 1.7230875492095947, + "learning_rate": 0.00015024518321170703, + "loss": 0.2471, + "step": 19330 + }, + { + "epoch": 0.7467469786478242, + "grad_norm": 1.73224937915802, + "learning_rate": 0.00015021944219210524, + "loss": 0.4877, + "step": 19340 + }, + { + "epoch": 0.747133093941851, + "grad_norm": 0.9023095369338989, + "learning_rate": 0.00015019370117250343, + "loss": 0.352, + "step": 19350 + }, + { + "epoch": 0.7475192092358778, + "grad_norm": 1.898983120918274, + "learning_rate": 0.00015016796015290167, + "loss": 0.3981, + "step": 19360 + }, + { + "epoch": 0.7479053245299047, + "grad_norm": 1.8544923067092896, + "learning_rate": 0.00015014221913329986, + "loss": 0.3015, + "step": 19370 + }, + { + "epoch": 0.7482914398239314, + "grad_norm": 1.1380795240402222, + "learning_rate": 0.0001501164781136981, + "loss": 0.3173, + "step": 19380 + }, + { + "epoch": 0.7486775551179582, + "grad_norm": 0.5071800351142883, + "learning_rate": 0.0001500907370940963, + "loss": 0.46, + "step": 19390 + }, + { + "epoch": 0.7490636704119851, + "grad_norm": 0.30739355087280273, + "learning_rate": 0.00015006499607449452, + "loss": 0.4942, + "step": 19400 + }, + { + "epoch": 0.7494497857060118, + "grad_norm": 1.1223585605621338, + "learning_rate": 0.00015003925505489274, + "loss": 0.3118, + "step": 19410 + }, + { + "epoch": 0.7498359010000386, + "grad_norm": 1.019545555114746, + "learning_rate": 0.00015001351403529092, + "loss": 0.3115, + "step": 19420 + }, + { + "epoch": 0.7502220162940654, + "grad_norm": 0.4567502439022064, + "learning_rate": 0.00014998777301568916, + "loss": 0.2932, + "step": 19430 + }, + { + "epoch": 0.7506081315880923, + "grad_norm": 1.669258952140808, + "learning_rate": 0.00014996203199608738, + "loss": 0.2889, + "step": 19440 + }, + { + "epoch": 0.750994246882119, + "grad_norm": 0.2787584364414215, + "learning_rate": 0.0001499362909764856, + "loss": 0.3482, + "step": 19450 + }, + { + "epoch": 0.7513803621761458, + "grad_norm": 1.6648303270339966, + "learning_rate": 0.0001499105499568838, + "loss": 0.3136, + "step": 19460 + }, + { + "epoch": 0.7517664774701726, + "grad_norm": 2.6357266902923584, + "learning_rate": 0.00014988480893728202, + "loss": 0.4643, + "step": 19470 + }, + { + "epoch": 0.7521525927641994, + "grad_norm": 1.8017394542694092, + "learning_rate": 0.00014985906791768023, + "loss": 0.469, + "step": 19480 + }, + { + "epoch": 0.7525387080582262, + "grad_norm": 1.8467847108840942, + "learning_rate": 0.00014983332689807842, + "loss": 0.1495, + "step": 19490 + }, + { + "epoch": 0.752924823352253, + "grad_norm": 2.9446980953216553, + "learning_rate": 0.00014980758587847666, + "loss": 0.3162, + "step": 19500 + }, + { + "epoch": 0.7533109386462797, + "grad_norm": 1.4076721668243408, + "learning_rate": 0.00014978184485887487, + "loss": 0.2356, + "step": 19510 + }, + { + "epoch": 0.7536970539403066, + "grad_norm": 1.2669463157653809, + "learning_rate": 0.00014975610383927308, + "loss": 0.3056, + "step": 19520 + }, + { + "epoch": 0.7540831692343334, + "grad_norm": 1.1390401124954224, + "learning_rate": 0.0001497303628196713, + "loss": 0.2785, + "step": 19530 + }, + { + "epoch": 0.7544692845283602, + "grad_norm": 3.095099925994873, + "learning_rate": 0.0001497046218000695, + "loss": 0.489, + "step": 19540 + }, + { + "epoch": 0.7548553998223869, + "grad_norm": 1.5737907886505127, + "learning_rate": 0.00014967888078046772, + "loss": 0.3981, + "step": 19550 + }, + { + "epoch": 0.7552415151164138, + "grad_norm": 1.4953045845031738, + "learning_rate": 0.00014965313976086594, + "loss": 0.3153, + "step": 19560 + }, + { + "epoch": 0.7556276304104406, + "grad_norm": 0.7709154486656189, + "learning_rate": 0.00014962739874126415, + "loss": 0.2904, + "step": 19570 + }, + { + "epoch": 0.7560137457044673, + "grad_norm": 2.118950366973877, + "learning_rate": 0.00014960165772166236, + "loss": 0.3736, + "step": 19580 + }, + { + "epoch": 0.7563998609984941, + "grad_norm": 2.9463138580322266, + "learning_rate": 0.00014957591670206058, + "loss": 0.3186, + "step": 19590 + }, + { + "epoch": 0.756785976292521, + "grad_norm": 1.3784689903259277, + "learning_rate": 0.0001495501756824588, + "loss": 0.3589, + "step": 19600 + }, + { + "epoch": 0.7571720915865477, + "grad_norm": 2.35467267036438, + "learning_rate": 0.000149524434662857, + "loss": 0.1282, + "step": 19610 + }, + { + "epoch": 0.7575582068805745, + "grad_norm": 0.7167999148368835, + "learning_rate": 0.00014949869364325522, + "loss": 0.3775, + "step": 19620 + }, + { + "epoch": 0.7579443221746013, + "grad_norm": 1.242785096168518, + "learning_rate": 0.00014947295262365343, + "loss": 0.3814, + "step": 19630 + }, + { + "epoch": 0.7583304374686282, + "grad_norm": 2.7875797748565674, + "learning_rate": 0.00014944721160405164, + "loss": 0.4764, + "step": 19640 + }, + { + "epoch": 0.7587165527626549, + "grad_norm": 1.2169462442398071, + "learning_rate": 0.00014942147058444986, + "loss": 0.3132, + "step": 19650 + }, + { + "epoch": 0.7591026680566817, + "grad_norm": 3.146204948425293, + "learning_rate": 0.00014939572956484807, + "loss": 0.248, + "step": 19660 + }, + { + "epoch": 0.7594887833507086, + "grad_norm": 2.2177276611328125, + "learning_rate": 0.00014936998854524628, + "loss": 0.3755, + "step": 19670 + }, + { + "epoch": 0.7598748986447353, + "grad_norm": 0.8063843250274658, + "learning_rate": 0.0001493442475256445, + "loss": 0.6341, + "step": 19680 + }, + { + "epoch": 0.7602610139387621, + "grad_norm": 0.18064215779304504, + "learning_rate": 0.0001493185065060427, + "loss": 0.3565, + "step": 19690 + }, + { + "epoch": 0.7606471292327889, + "grad_norm": 1.4401954412460327, + "learning_rate": 0.00014929276548644092, + "loss": 0.1873, + "step": 19700 + }, + { + "epoch": 0.7610332445268158, + "grad_norm": 0.7269515991210938, + "learning_rate": 0.00014926702446683914, + "loss": 0.2765, + "step": 19710 + }, + { + "epoch": 0.7614193598208425, + "grad_norm": 1.2779995203018188, + "learning_rate": 0.00014924128344723735, + "loss": 0.3633, + "step": 19720 + }, + { + "epoch": 0.7618054751148693, + "grad_norm": 1.3330426216125488, + "learning_rate": 0.00014921554242763556, + "loss": 0.3736, + "step": 19730 + }, + { + "epoch": 0.7621915904088961, + "grad_norm": 1.4269347190856934, + "learning_rate": 0.00014918980140803378, + "loss": 0.2572, + "step": 19740 + }, + { + "epoch": 0.7625777057029229, + "grad_norm": 0.9075976610183716, + "learning_rate": 0.000149164060388432, + "loss": 0.3609, + "step": 19750 + }, + { + "epoch": 0.7629638209969497, + "grad_norm": 1.7566559314727783, + "learning_rate": 0.0001491383193688302, + "loss": 0.3134, + "step": 19760 + }, + { + "epoch": 0.7633499362909765, + "grad_norm": 2.4620914459228516, + "learning_rate": 0.00014911257834922842, + "loss": 0.3559, + "step": 19770 + }, + { + "epoch": 0.7637360515850032, + "grad_norm": 0.9474597573280334, + "learning_rate": 0.00014908683732962663, + "loss": 0.3821, + "step": 19780 + }, + { + "epoch": 0.7641221668790301, + "grad_norm": 0.6298363208770752, + "learning_rate": 0.00014906109631002484, + "loss": 0.2801, + "step": 19790 + }, + { + "epoch": 0.7645082821730569, + "grad_norm": 0.213288351893425, + "learning_rate": 0.00014903535529042306, + "loss": 0.368, + "step": 19800 + }, + { + "epoch": 0.7648943974670837, + "grad_norm": 1.4412375688552856, + "learning_rate": 0.00014900961427082127, + "loss": 0.3972, + "step": 19810 + }, + { + "epoch": 0.7652805127611104, + "grad_norm": 1.6808812618255615, + "learning_rate": 0.00014898387325121948, + "loss": 0.5302, + "step": 19820 + }, + { + "epoch": 0.7656666280551373, + "grad_norm": 0.922726571559906, + "learning_rate": 0.0001489581322316177, + "loss": 0.3299, + "step": 19830 + }, + { + "epoch": 0.7660527433491641, + "grad_norm": 0.5058152079582214, + "learning_rate": 0.0001489323912120159, + "loss": 0.4485, + "step": 19840 + }, + { + "epoch": 0.7664388586431908, + "grad_norm": 1.3025776147842407, + "learning_rate": 0.00014890665019241412, + "loss": 0.266, + "step": 19850 + }, + { + "epoch": 0.7668249739372176, + "grad_norm": 2.0282516479492188, + "learning_rate": 0.00014888090917281234, + "loss": 0.7752, + "step": 19860 + }, + { + "epoch": 0.7672110892312445, + "grad_norm": 0.6557582020759583, + "learning_rate": 0.00014885516815321058, + "loss": 0.256, + "step": 19870 + }, + { + "epoch": 0.7675972045252712, + "grad_norm": 0.13688494265079498, + "learning_rate": 0.00014882942713360876, + "loss": 0.263, + "step": 19880 + }, + { + "epoch": 0.767983319819298, + "grad_norm": 0.04091642051935196, + "learning_rate": 0.00014880368611400698, + "loss": 0.276, + "step": 19890 + }, + { + "epoch": 0.7683694351133248, + "grad_norm": 3.054969072341919, + "learning_rate": 0.0001487779450944052, + "loss": 0.5198, + "step": 19900 + }, + { + "epoch": 0.7687555504073517, + "grad_norm": 0.5707372426986694, + "learning_rate": 0.0001487522040748034, + "loss": 0.2715, + "step": 19910 + }, + { + "epoch": 0.7691416657013784, + "grad_norm": 0.477830708026886, + "learning_rate": 0.00014872646305520162, + "loss": 0.2517, + "step": 19920 + }, + { + "epoch": 0.7695277809954052, + "grad_norm": 1.0550785064697266, + "learning_rate": 0.00014870072203559983, + "loss": 0.2552, + "step": 19930 + }, + { + "epoch": 0.7699138962894321, + "grad_norm": 1.9678715467453003, + "learning_rate": 0.00014867498101599807, + "loss": 0.4487, + "step": 19940 + }, + { + "epoch": 0.7703000115834588, + "grad_norm": 0.4506283104419708, + "learning_rate": 0.00014864923999639626, + "loss": 0.2632, + "step": 19950 + }, + { + "epoch": 0.7706861268774856, + "grad_norm": 1.8445035219192505, + "learning_rate": 0.0001486234989767945, + "loss": 0.3315, + "step": 19960 + }, + { + "epoch": 0.7710722421715124, + "grad_norm": 2.6394076347351074, + "learning_rate": 0.00014859775795719268, + "loss": 0.3216, + "step": 19970 + }, + { + "epoch": 0.7714583574655393, + "grad_norm": 0.6717782020568848, + "learning_rate": 0.0001485720169375909, + "loss": 0.3531, + "step": 19980 + }, + { + "epoch": 0.771844472759566, + "grad_norm": 0.5959204435348511, + "learning_rate": 0.0001485462759179891, + "loss": 0.2095, + "step": 19990 + }, + { + "epoch": 0.7722305880535928, + "grad_norm": 1.874375581741333, + "learning_rate": 0.00014852053489838732, + "loss": 0.4845, + "step": 20000 + }, + { + "epoch": 0.7726167033476196, + "grad_norm": 2.6810474395751953, + "learning_rate": 0.00014849479387878556, + "loss": 0.471, + "step": 20010 + }, + { + "epoch": 0.7730028186416464, + "grad_norm": 0.5498594045639038, + "learning_rate": 0.00014846905285918375, + "loss": 0.2265, + "step": 20020 + }, + { + "epoch": 0.7733889339356732, + "grad_norm": 0.7658601999282837, + "learning_rate": 0.000148443311839582, + "loss": 0.4459, + "step": 20030 + }, + { + "epoch": 0.7737750492297, + "grad_norm": 0.5597706437110901, + "learning_rate": 0.00014841757081998018, + "loss": 0.4317, + "step": 20040 + }, + { + "epoch": 0.7741611645237267, + "grad_norm": 0.740342915058136, + "learning_rate": 0.0001483918298003784, + "loss": 0.4158, + "step": 20050 + }, + { + "epoch": 0.7745472798177536, + "grad_norm": 0.6069484949111938, + "learning_rate": 0.00014836608878077663, + "loss": 0.2363, + "step": 20060 + }, + { + "epoch": 0.7749333951117804, + "grad_norm": 0.9331381916999817, + "learning_rate": 0.00014834034776117482, + "loss": 0.2063, + "step": 20070 + }, + { + "epoch": 0.7753195104058072, + "grad_norm": 1.298399806022644, + "learning_rate": 0.00014831460674157306, + "loss": 0.319, + "step": 20080 + }, + { + "epoch": 0.7757056256998339, + "grad_norm": 1.887229323387146, + "learning_rate": 0.00014828886572197124, + "loss": 0.3805, + "step": 20090 + }, + { + "epoch": 0.7760917409938608, + "grad_norm": 0.5734463930130005, + "learning_rate": 0.00014826312470236948, + "loss": 0.392, + "step": 20100 + }, + { + "epoch": 0.7764778562878876, + "grad_norm": 1.10410475730896, + "learning_rate": 0.00014823738368276767, + "loss": 0.3222, + "step": 20110 + }, + { + "epoch": 0.7768639715819143, + "grad_norm": 3.1354923248291016, + "learning_rate": 0.00014821164266316588, + "loss": 0.3434, + "step": 20120 + }, + { + "epoch": 0.7772500868759411, + "grad_norm": 1.9921913146972656, + "learning_rate": 0.00014818590164356412, + "loss": 0.222, + "step": 20130 + }, + { + "epoch": 0.777636202169968, + "grad_norm": 0.38615912199020386, + "learning_rate": 0.0001481601606239623, + "loss": 0.1859, + "step": 20140 + }, + { + "epoch": 0.7780223174639948, + "grad_norm": 0.919252336025238, + "learning_rate": 0.00014813441960436055, + "loss": 0.3889, + "step": 20150 + }, + { + "epoch": 0.7784084327580215, + "grad_norm": 0.9474624991416931, + "learning_rate": 0.00014810867858475874, + "loss": 0.3086, + "step": 20160 + }, + { + "epoch": 0.7787945480520484, + "grad_norm": 1.1992007493972778, + "learning_rate": 0.00014808293756515698, + "loss": 0.3949, + "step": 20170 + }, + { + "epoch": 0.7791806633460752, + "grad_norm": 2.2405364513397217, + "learning_rate": 0.00014805719654555516, + "loss": 0.4963, + "step": 20180 + }, + { + "epoch": 0.7795667786401019, + "grad_norm": 0.9163286685943604, + "learning_rate": 0.00014803145552595338, + "loss": 0.4022, + "step": 20190 + }, + { + "epoch": 0.7799528939341287, + "grad_norm": 0.995309591293335, + "learning_rate": 0.00014800571450635162, + "loss": 0.244, + "step": 20200 + }, + { + "epoch": 0.7803390092281556, + "grad_norm": 3.0966508388519287, + "learning_rate": 0.0001479799734867498, + "loss": 0.4165, + "step": 20210 + }, + { + "epoch": 0.7807251245221823, + "grad_norm": 1.7679264545440674, + "learning_rate": 0.00014795423246714804, + "loss": 0.3264, + "step": 20220 + }, + { + "epoch": 0.7811112398162091, + "grad_norm": 3.405724048614502, + "learning_rate": 0.00014792849144754623, + "loss": 0.5292, + "step": 20230 + }, + { + "epoch": 0.7814973551102359, + "grad_norm": 0.5335774421691895, + "learning_rate": 0.00014790275042794447, + "loss": 0.2901, + "step": 20240 + }, + { + "epoch": 0.7818834704042628, + "grad_norm": 2.332176446914673, + "learning_rate": 0.00014787700940834268, + "loss": 0.3116, + "step": 20250 + }, + { + "epoch": 0.7822695856982895, + "grad_norm": 1.226844072341919, + "learning_rate": 0.00014785126838874087, + "loss": 0.332, + "step": 20260 + }, + { + "epoch": 0.7826557009923163, + "grad_norm": 0.8310544490814209, + "learning_rate": 0.0001478255273691391, + "loss": 0.3929, + "step": 20270 + }, + { + "epoch": 0.7830418162863431, + "grad_norm": 4.065925121307373, + "learning_rate": 0.0001477997863495373, + "loss": 0.3044, + "step": 20280 + }, + { + "epoch": 0.7834279315803699, + "grad_norm": 1.3661054372787476, + "learning_rate": 0.00014777404532993554, + "loss": 0.2396, + "step": 20290 + }, + { + "epoch": 0.7838140468743967, + "grad_norm": 2.047938346862793, + "learning_rate": 0.00014774830431033372, + "loss": 0.2954, + "step": 20300 + }, + { + "epoch": 0.7842001621684235, + "grad_norm": 2.099759817123413, + "learning_rate": 0.00014772256329073196, + "loss": 0.1987, + "step": 20310 + }, + { + "epoch": 0.7845862774624502, + "grad_norm": 4.439823627471924, + "learning_rate": 0.00014769682227113018, + "loss": 0.2449, + "step": 20320 + }, + { + "epoch": 0.7849723927564771, + "grad_norm": 1.894250750541687, + "learning_rate": 0.00014767108125152836, + "loss": 0.3335, + "step": 20330 + }, + { + "epoch": 0.7853585080505039, + "grad_norm": 2.2916924953460693, + "learning_rate": 0.0001476453402319266, + "loss": 0.3828, + "step": 20340 + }, + { + "epoch": 0.7857446233445307, + "grad_norm": 3.7847397327423096, + "learning_rate": 0.0001476195992123248, + "loss": 0.2142, + "step": 20350 + }, + { + "epoch": 0.7861307386385574, + "grad_norm": 2.7507593631744385, + "learning_rate": 0.00014759385819272303, + "loss": 0.39, + "step": 20360 + }, + { + "epoch": 0.7865168539325843, + "grad_norm": 2.500195264816284, + "learning_rate": 0.00014756811717312124, + "loss": 0.4163, + "step": 20370 + }, + { + "epoch": 0.7869029692266111, + "grad_norm": 1.3699760437011719, + "learning_rate": 0.00014754237615351946, + "loss": 0.2407, + "step": 20380 + }, + { + "epoch": 0.7872890845206378, + "grad_norm": 0.9736176133155823, + "learning_rate": 0.00014751663513391767, + "loss": 0.3644, + "step": 20390 + }, + { + "epoch": 0.7876751998146646, + "grad_norm": 0.9103217720985413, + "learning_rate": 0.00014749089411431588, + "loss": 0.1304, + "step": 20400 + }, + { + "epoch": 0.7880613151086915, + "grad_norm": 0.7736025452613831, + "learning_rate": 0.0001474651530947141, + "loss": 0.3848, + "step": 20410 + }, + { + "epoch": 0.7884474304027183, + "grad_norm": 1.6167576313018799, + "learning_rate": 0.00014743941207511228, + "loss": 0.3469, + "step": 20420 + }, + { + "epoch": 0.788833545696745, + "grad_norm": 0.1359117031097412, + "learning_rate": 0.00014741367105551052, + "loss": 0.28, + "step": 20430 + }, + { + "epoch": 0.7892196609907719, + "grad_norm": 0.20857305824756622, + "learning_rate": 0.00014738793003590874, + "loss": 0.2406, + "step": 20440 + }, + { + "epoch": 0.7896057762847987, + "grad_norm": 0.9618992805480957, + "learning_rate": 0.00014736218901630695, + "loss": 0.243, + "step": 20450 + }, + { + "epoch": 0.7899918915788254, + "grad_norm": 0.4663112163543701, + "learning_rate": 0.00014733644799670516, + "loss": 0.214, + "step": 20460 + }, + { + "epoch": 0.7903780068728522, + "grad_norm": 1.3922615051269531, + "learning_rate": 0.00014731070697710338, + "loss": 0.4601, + "step": 20470 + }, + { + "epoch": 0.7907641221668791, + "grad_norm": 1.5706521272659302, + "learning_rate": 0.0001472849659575016, + "loss": 0.2187, + "step": 20480 + }, + { + "epoch": 0.7911502374609058, + "grad_norm": 1.9315848350524902, + "learning_rate": 0.00014725922493789978, + "loss": 0.2691, + "step": 20490 + }, + { + "epoch": 0.7915363527549326, + "grad_norm": 2.6131045818328857, + "learning_rate": 0.00014723348391829802, + "loss": 0.335, + "step": 20500 + }, + { + "epoch": 0.7919224680489594, + "grad_norm": 2.6561343669891357, + "learning_rate": 0.00014720774289869623, + "loss": 0.2166, + "step": 20510 + }, + { + "epoch": 0.7923085833429863, + "grad_norm": 1.247574806213379, + "learning_rate": 0.00014718200187909444, + "loss": 0.4276, + "step": 20520 + }, + { + "epoch": 0.792694698637013, + "grad_norm": 0.7353020310401917, + "learning_rate": 0.00014715626085949266, + "loss": 0.3442, + "step": 20530 + }, + { + "epoch": 0.7930808139310398, + "grad_norm": 0.1100919172167778, + "learning_rate": 0.00014713051983989087, + "loss": 0.21, + "step": 20540 + }, + { + "epoch": 0.7934669292250666, + "grad_norm": 0.6608699560165405, + "learning_rate": 0.00014710477882028908, + "loss": 0.2593, + "step": 20550 + }, + { + "epoch": 0.7938530445190934, + "grad_norm": 1.0959700345993042, + "learning_rate": 0.0001470790378006873, + "loss": 0.1864, + "step": 20560 + }, + { + "epoch": 0.7942391598131202, + "grad_norm": 1.469408392906189, + "learning_rate": 0.0001470532967810855, + "loss": 0.251, + "step": 20570 + }, + { + "epoch": 0.794625275107147, + "grad_norm": 1.4239304065704346, + "learning_rate": 0.00014702755576148372, + "loss": 0.1782, + "step": 20580 + }, + { + "epoch": 0.7950113904011737, + "grad_norm": 0.5389681458473206, + "learning_rate": 0.00014700181474188194, + "loss": 0.3228, + "step": 20590 + }, + { + "epoch": 0.7953975056952006, + "grad_norm": 0.4744633436203003, + "learning_rate": 0.00014697607372228015, + "loss": 0.5448, + "step": 20600 + }, + { + "epoch": 0.7957836209892274, + "grad_norm": 1.1155211925506592, + "learning_rate": 0.00014695033270267836, + "loss": 0.3341, + "step": 20610 + }, + { + "epoch": 0.7961697362832542, + "grad_norm": 1.4301745891571045, + "learning_rate": 0.00014692459168307658, + "loss": 0.2376, + "step": 20620 + }, + { + "epoch": 0.7965558515772809, + "grad_norm": 2.0889878273010254, + "learning_rate": 0.0001468988506634748, + "loss": 0.252, + "step": 20630 + }, + { + "epoch": 0.7969419668713078, + "grad_norm": 1.420873761177063, + "learning_rate": 0.000146873109643873, + "loss": 0.4033, + "step": 20640 + }, + { + "epoch": 0.7973280821653346, + "grad_norm": 1.9909567832946777, + "learning_rate": 0.00014684736862427122, + "loss": 0.5206, + "step": 20650 + }, + { + "epoch": 0.7977141974593613, + "grad_norm": 1.3584216833114624, + "learning_rate": 0.00014682162760466943, + "loss": 0.2643, + "step": 20660 + }, + { + "epoch": 0.7981003127533881, + "grad_norm": 1.5488578081130981, + "learning_rate": 0.00014679588658506764, + "loss": 0.4098, + "step": 20670 + }, + { + "epoch": 0.798486428047415, + "grad_norm": 2.0128777027130127, + "learning_rate": 0.00014677014556546586, + "loss": 0.2498, + "step": 20680 + }, + { + "epoch": 0.7988725433414418, + "grad_norm": 0.6591671705245972, + "learning_rate": 0.00014674440454586407, + "loss": 0.3131, + "step": 20690 + }, + { + "epoch": 0.7992586586354685, + "grad_norm": 0.8321843147277832, + "learning_rate": 0.00014671866352626228, + "loss": 0.3849, + "step": 20700 + }, + { + "epoch": 0.7996447739294954, + "grad_norm": 1.4672768115997314, + "learning_rate": 0.0001466929225066605, + "loss": 0.3056, + "step": 20710 + }, + { + "epoch": 0.8000308892235222, + "grad_norm": 1.5972867012023926, + "learning_rate": 0.0001466671814870587, + "loss": 0.2699, + "step": 20720 + }, + { + "epoch": 0.8004170045175489, + "grad_norm": 0.685972273349762, + "learning_rate": 0.00014664144046745692, + "loss": 0.2598, + "step": 20730 + }, + { + "epoch": 0.8008031198115757, + "grad_norm": 1.2639611959457397, + "learning_rate": 0.00014661569944785514, + "loss": 0.2158, + "step": 20740 + }, + { + "epoch": 0.8011892351056026, + "grad_norm": 2.0752620697021484, + "learning_rate": 0.00014658995842825335, + "loss": 0.2782, + "step": 20750 + }, + { + "epoch": 0.8015753503996293, + "grad_norm": 4.334362983703613, + "learning_rate": 0.00014656421740865156, + "loss": 0.5083, + "step": 20760 + }, + { + "epoch": 0.8019614656936561, + "grad_norm": 2.233436107635498, + "learning_rate": 0.00014653847638904978, + "loss": 0.4378, + "step": 20770 + }, + { + "epoch": 0.8023475809876829, + "grad_norm": 0.5392621755599976, + "learning_rate": 0.000146512735369448, + "loss": 0.4467, + "step": 20780 + }, + { + "epoch": 0.8027336962817098, + "grad_norm": 0.3768058121204376, + "learning_rate": 0.0001464869943498462, + "loss": 0.3542, + "step": 20790 + }, + { + "epoch": 0.8031198115757365, + "grad_norm": 0.5712292194366455, + "learning_rate": 0.00014646125333024442, + "loss": 0.281, + "step": 20800 + }, + { + "epoch": 0.8035059268697633, + "grad_norm": 0.8674315214157104, + "learning_rate": 0.00014643551231064263, + "loss": 0.2945, + "step": 20810 + }, + { + "epoch": 0.8038920421637901, + "grad_norm": 1.6206015348434448, + "learning_rate": 0.00014640977129104084, + "loss": 0.2627, + "step": 20820 + }, + { + "epoch": 0.804278157457817, + "grad_norm": 1.8807138204574585, + "learning_rate": 0.00014638403027143906, + "loss": 0.3768, + "step": 20830 + }, + { + "epoch": 0.8046642727518437, + "grad_norm": 0.9721212387084961, + "learning_rate": 0.00014635828925183727, + "loss": 0.4511, + "step": 20840 + }, + { + "epoch": 0.8050503880458705, + "grad_norm": 0.569038987159729, + "learning_rate": 0.00014633254823223548, + "loss": 0.2235, + "step": 20850 + }, + { + "epoch": 0.8054365033398972, + "grad_norm": 0.5981199741363525, + "learning_rate": 0.0001463068072126337, + "loss": 0.1413, + "step": 20860 + }, + { + "epoch": 0.8058226186339241, + "grad_norm": 3.696936845779419, + "learning_rate": 0.00014628106619303194, + "loss": 0.3779, + "step": 20870 + }, + { + "epoch": 0.8062087339279509, + "grad_norm": 1.5039314031600952, + "learning_rate": 0.00014625532517343012, + "loss": 0.2117, + "step": 20880 + }, + { + "epoch": 0.8065948492219777, + "grad_norm": 1.7800476551055908, + "learning_rate": 0.00014622958415382834, + "loss": 0.2343, + "step": 20890 + }, + { + "epoch": 0.8069809645160044, + "grad_norm": 0.18587611615657806, + "learning_rate": 0.00014620384313422655, + "loss": 0.2229, + "step": 20900 + }, + { + "epoch": 0.8073670798100313, + "grad_norm": 3.5351600646972656, + "learning_rate": 0.00014617810211462476, + "loss": 0.3632, + "step": 20910 + }, + { + "epoch": 0.8077531951040581, + "grad_norm": 3.9036381244659424, + "learning_rate": 0.00014615236109502298, + "loss": 0.2941, + "step": 20920 + }, + { + "epoch": 0.8081393103980848, + "grad_norm": 1.9554537534713745, + "learning_rate": 0.0001461266200754212, + "loss": 0.2685, + "step": 20930 + }, + { + "epoch": 0.8085254256921116, + "grad_norm": 1.0424940586090088, + "learning_rate": 0.00014610087905581943, + "loss": 0.2617, + "step": 20940 + }, + { + "epoch": 0.8089115409861385, + "grad_norm": 6.593061923980713, + "learning_rate": 0.00014607513803621762, + "loss": 0.5034, + "step": 20950 + }, + { + "epoch": 0.8092976562801653, + "grad_norm": 11.373255729675293, + "learning_rate": 0.00014604939701661583, + "loss": 0.2694, + "step": 20960 + }, + { + "epoch": 0.809683771574192, + "grad_norm": 1.4843833446502686, + "learning_rate": 0.00014602365599701404, + "loss": 0.4526, + "step": 20970 + }, + { + "epoch": 0.8100698868682189, + "grad_norm": 3.6086366176605225, + "learning_rate": 0.00014599791497741225, + "loss": 0.4536, + "step": 20980 + }, + { + "epoch": 0.8104560021622457, + "grad_norm": 0.6381124258041382, + "learning_rate": 0.00014597217395781047, + "loss": 0.2918, + "step": 20990 + }, + { + "epoch": 0.8108421174562724, + "grad_norm": 1.9507087469100952, + "learning_rate": 0.00014594643293820868, + "loss": 0.1625, + "step": 21000 + }, + { + "epoch": 0.8112282327502992, + "grad_norm": 0.954914391040802, + "learning_rate": 0.00014592069191860692, + "loss": 0.4204, + "step": 21010 + }, + { + "epoch": 0.8116143480443261, + "grad_norm": 2.633601427078247, + "learning_rate": 0.0001458949508990051, + "loss": 0.4017, + "step": 21020 + }, + { + "epoch": 0.8120004633383529, + "grad_norm": 0.7492280602455139, + "learning_rate": 0.00014586920987940332, + "loss": 0.3209, + "step": 21030 + }, + { + "epoch": 0.8123865786323796, + "grad_norm": 1.1670303344726562, + "learning_rate": 0.00014584346885980153, + "loss": 0.3361, + "step": 21040 + }, + { + "epoch": 0.8127726939264064, + "grad_norm": 0.8462283611297607, + "learning_rate": 0.00014581772784019975, + "loss": 0.3156, + "step": 21050 + }, + { + "epoch": 0.8131588092204333, + "grad_norm": 2.151671886444092, + "learning_rate": 0.000145791986820598, + "loss": 0.2081, + "step": 21060 + }, + { + "epoch": 0.81354492451446, + "grad_norm": 1.0742170810699463, + "learning_rate": 0.00014576624580099617, + "loss": 0.2247, + "step": 21070 + }, + { + "epoch": 0.8139310398084868, + "grad_norm": 1.2256931066513062, + "learning_rate": 0.00014574050478139442, + "loss": 0.3542, + "step": 21080 + }, + { + "epoch": 0.8143171551025136, + "grad_norm": 3.740055561065674, + "learning_rate": 0.0001457147637617926, + "loss": 0.1856, + "step": 21090 + }, + { + "epoch": 0.8147032703965404, + "grad_norm": 0.03290783613920212, + "learning_rate": 0.00014568902274219081, + "loss": 0.2616, + "step": 21100 + }, + { + "epoch": 0.8150893856905672, + "grad_norm": 0.13995541632175446, + "learning_rate": 0.00014566328172258903, + "loss": 0.2107, + "step": 21110 + }, + { + "epoch": 0.815475500984594, + "grad_norm": 0.44371533393859863, + "learning_rate": 0.00014563754070298724, + "loss": 0.3091, + "step": 21120 + }, + { + "epoch": 0.8158616162786207, + "grad_norm": 2.7269155979156494, + "learning_rate": 0.00014561179968338548, + "loss": 0.5674, + "step": 21130 + }, + { + "epoch": 0.8162477315726476, + "grad_norm": 0.7148515582084656, + "learning_rate": 0.00014558605866378367, + "loss": 0.4721, + "step": 21140 + }, + { + "epoch": 0.8166338468666744, + "grad_norm": 1.0939961671829224, + "learning_rate": 0.0001455603176441819, + "loss": 0.2666, + "step": 21150 + }, + { + "epoch": 0.8170199621607012, + "grad_norm": 1.7923939228057861, + "learning_rate": 0.0001455345766245801, + "loss": 0.2634, + "step": 21160 + }, + { + "epoch": 0.8174060774547279, + "grad_norm": 0.725130021572113, + "learning_rate": 0.00014550883560497834, + "loss": 0.1968, + "step": 21170 + }, + { + "epoch": 0.8177921927487548, + "grad_norm": 0.443892240524292, + "learning_rate": 0.00014548309458537655, + "loss": 0.556, + "step": 21180 + }, + { + "epoch": 0.8181783080427816, + "grad_norm": 1.3551362752914429, + "learning_rate": 0.00014545735356577473, + "loss": 0.4115, + "step": 21190 + }, + { + "epoch": 0.8185644233368083, + "grad_norm": 0.6360037922859192, + "learning_rate": 0.00014543161254617297, + "loss": 0.3176, + "step": 21200 + }, + { + "epoch": 0.8189505386308351, + "grad_norm": 2.634549140930176, + "learning_rate": 0.00014540587152657116, + "loss": 0.3662, + "step": 21210 + }, + { + "epoch": 0.819336653924862, + "grad_norm": 3.267479181289673, + "learning_rate": 0.0001453801305069694, + "loss": 0.2925, + "step": 21220 + }, + { + "epoch": 0.8197227692188888, + "grad_norm": 1.3607991933822632, + "learning_rate": 0.0001453543894873676, + "loss": 0.1795, + "step": 21230 + }, + { + "epoch": 0.8201088845129155, + "grad_norm": 0.6499636769294739, + "learning_rate": 0.00014532864846776583, + "loss": 0.4451, + "step": 21240 + }, + { + "epoch": 0.8204949998069424, + "grad_norm": 0.7486141920089722, + "learning_rate": 0.00014530290744816404, + "loss": 0.3982, + "step": 21250 + }, + { + "epoch": 0.8208811151009692, + "grad_norm": 0.6481244564056396, + "learning_rate": 0.00014527716642856223, + "loss": 0.3358, + "step": 21260 + }, + { + "epoch": 0.8212672303949959, + "grad_norm": 1.0736982822418213, + "learning_rate": 0.00014525142540896047, + "loss": 0.5264, + "step": 21270 + }, + { + "epoch": 0.8216533456890227, + "grad_norm": 2.0467801094055176, + "learning_rate": 0.00014522568438935865, + "loss": 0.3645, + "step": 21280 + }, + { + "epoch": 0.8220394609830496, + "grad_norm": 2.671499013900757, + "learning_rate": 0.0001451999433697569, + "loss": 0.4779, + "step": 21290 + }, + { + "epoch": 0.8224255762770764, + "grad_norm": 1.4449695348739624, + "learning_rate": 0.00014517420235015508, + "loss": 0.3555, + "step": 21300 + }, + { + "epoch": 0.8228116915711031, + "grad_norm": 1.7484570741653442, + "learning_rate": 0.00014514846133055332, + "loss": 0.2921, + "step": 21310 + }, + { + "epoch": 0.8231978068651299, + "grad_norm": 0.9985783100128174, + "learning_rate": 0.00014512272031095153, + "loss": 0.1861, + "step": 21320 + }, + { + "epoch": 0.8235839221591568, + "grad_norm": 2.0824766159057617, + "learning_rate": 0.00014509697929134972, + "loss": 0.3582, + "step": 21330 + }, + { + "epoch": 0.8239700374531835, + "grad_norm": 0.8448216915130615, + "learning_rate": 0.00014507123827174796, + "loss": 0.3674, + "step": 21340 + }, + { + "epoch": 0.8243561527472103, + "grad_norm": 2.027111053466797, + "learning_rate": 0.00014504549725214615, + "loss": 0.3297, + "step": 21350 + }, + { + "epoch": 0.8247422680412371, + "grad_norm": 1.560604214668274, + "learning_rate": 0.0001450197562325444, + "loss": 0.3303, + "step": 21360 + }, + { + "epoch": 0.825128383335264, + "grad_norm": 2.179563045501709, + "learning_rate": 0.0001449940152129426, + "loss": 0.1704, + "step": 21370 + }, + { + "epoch": 0.8255144986292907, + "grad_norm": 1.6268993616104126, + "learning_rate": 0.00014496827419334081, + "loss": 0.3316, + "step": 21380 + }, + { + "epoch": 0.8259006139233175, + "grad_norm": 0.8986232280731201, + "learning_rate": 0.00014494253317373903, + "loss": 0.3361, + "step": 21390 + }, + { + "epoch": 0.8262867292173443, + "grad_norm": 0.8017566204071045, + "learning_rate": 0.00014491679215413721, + "loss": 0.3992, + "step": 21400 + }, + { + "epoch": 0.8266728445113711, + "grad_norm": 0.879162073135376, + "learning_rate": 0.00014489105113453545, + "loss": 0.3747, + "step": 21410 + }, + { + "epoch": 0.8270589598053979, + "grad_norm": 1.948309302330017, + "learning_rate": 0.00014486531011493364, + "loss": 0.2323, + "step": 21420 + }, + { + "epoch": 0.8274450750994247, + "grad_norm": 1.34186851978302, + "learning_rate": 0.00014483956909533188, + "loss": 0.3856, + "step": 21430 + }, + { + "epoch": 0.8278311903934514, + "grad_norm": 1.3884105682373047, + "learning_rate": 0.0001448138280757301, + "loss": 0.3044, + "step": 21440 + }, + { + "epoch": 0.8282173056874783, + "grad_norm": 1.3283358812332153, + "learning_rate": 0.0001447880870561283, + "loss": 0.3624, + "step": 21450 + }, + { + "epoch": 0.8286034209815051, + "grad_norm": 13.829493522644043, + "learning_rate": 0.00014476234603652652, + "loss": 0.2217, + "step": 21460 + }, + { + "epoch": 0.8289895362755318, + "grad_norm": 3.4602255821228027, + "learning_rate": 0.0001447366050169247, + "loss": 0.4964, + "step": 21470 + }, + { + "epoch": 0.8293756515695587, + "grad_norm": 0.42417749762535095, + "learning_rate": 0.00014471086399732295, + "loss": 0.2262, + "step": 21480 + }, + { + "epoch": 0.8297617668635855, + "grad_norm": 3.1674726009368896, + "learning_rate": 0.00014468512297772113, + "loss": 0.3327, + "step": 21490 + }, + { + "epoch": 0.8301478821576123, + "grad_norm": 0.7226410508155823, + "learning_rate": 0.00014465938195811937, + "loss": 0.3152, + "step": 21500 + }, + { + "epoch": 0.830533997451639, + "grad_norm": 0.7477544546127319, + "learning_rate": 0.0001446336409385176, + "loss": 0.3676, + "step": 21510 + }, + { + "epoch": 0.8309201127456659, + "grad_norm": 1.6237748861312866, + "learning_rate": 0.0001446078999189158, + "loss": 0.336, + "step": 21520 + }, + { + "epoch": 0.8313062280396927, + "grad_norm": 2.8118655681610107, + "learning_rate": 0.00014458215889931401, + "loss": 0.3513, + "step": 21530 + }, + { + "epoch": 0.8316923433337194, + "grad_norm": 2.6571335792541504, + "learning_rate": 0.0001445564178797122, + "loss": 0.4008, + "step": 21540 + }, + { + "epoch": 0.8320784586277462, + "grad_norm": 0.7042214870452881, + "learning_rate": 0.00014453067686011044, + "loss": 0.3433, + "step": 21550 + }, + { + "epoch": 0.8324645739217731, + "grad_norm": 0.1551884561777115, + "learning_rate": 0.00014450493584050865, + "loss": 0.1748, + "step": 21560 + }, + { + "epoch": 0.8328506892157999, + "grad_norm": 1.2595586776733398, + "learning_rate": 0.00014447919482090687, + "loss": 0.2567, + "step": 21570 + }, + { + "epoch": 0.8332368045098266, + "grad_norm": 3.800837516784668, + "learning_rate": 0.00014445345380130508, + "loss": 0.31, + "step": 21580 + }, + { + "epoch": 0.8336229198038534, + "grad_norm": 2.2269585132598877, + "learning_rate": 0.0001444277127817033, + "loss": 0.5146, + "step": 21590 + }, + { + "epoch": 0.8340090350978803, + "grad_norm": 0.5263709425926208, + "learning_rate": 0.0001444019717621015, + "loss": 0.2369, + "step": 21600 + }, + { + "epoch": 0.834395150391907, + "grad_norm": 0.27564361691474915, + "learning_rate": 0.00014437623074249972, + "loss": 0.308, + "step": 21610 + }, + { + "epoch": 0.8347812656859338, + "grad_norm": 4.639162540435791, + "learning_rate": 0.00014435048972289793, + "loss": 0.3806, + "step": 21620 + }, + { + "epoch": 0.8351673809799606, + "grad_norm": 0.9607310891151428, + "learning_rate": 0.00014432474870329615, + "loss": 0.2245, + "step": 21630 + }, + { + "epoch": 0.8355534962739875, + "grad_norm": 4.01082706451416, + "learning_rate": 0.00014429900768369436, + "loss": 0.3934, + "step": 21640 + }, + { + "epoch": 0.8359396115680142, + "grad_norm": 0.9401382803916931, + "learning_rate": 0.00014427326666409257, + "loss": 0.4726, + "step": 21650 + }, + { + "epoch": 0.836325726862041, + "grad_norm": 2.1189887523651123, + "learning_rate": 0.0001442475256444908, + "loss": 0.5434, + "step": 21660 + }, + { + "epoch": 0.8367118421560678, + "grad_norm": 2.370849132537842, + "learning_rate": 0.000144221784624889, + "loss": 0.3885, + "step": 21670 + }, + { + "epoch": 0.8370979574500946, + "grad_norm": 0.595461368560791, + "learning_rate": 0.00014419604360528721, + "loss": 0.3811, + "step": 21680 + }, + { + "epoch": 0.8374840727441214, + "grad_norm": 0.9013121128082275, + "learning_rate": 0.00014417030258568543, + "loss": 0.2406, + "step": 21690 + }, + { + "epoch": 0.8378701880381482, + "grad_norm": 1.3803203105926514, + "learning_rate": 0.00014414456156608364, + "loss": 0.2197, + "step": 21700 + }, + { + "epoch": 0.8382563033321749, + "grad_norm": 1.6163750886917114, + "learning_rate": 0.00014411882054648185, + "loss": 0.2622, + "step": 21710 + }, + { + "epoch": 0.8386424186262018, + "grad_norm": 3.604384660720825, + "learning_rate": 0.00014409307952688007, + "loss": 0.3688, + "step": 21720 + }, + { + "epoch": 0.8390285339202286, + "grad_norm": 1.4415024518966675, + "learning_rate": 0.00014406733850727828, + "loss": 0.2375, + "step": 21730 + }, + { + "epoch": 0.8394146492142553, + "grad_norm": 1.4819844961166382, + "learning_rate": 0.0001440415974876765, + "loss": 0.4065, + "step": 21740 + }, + { + "epoch": 0.8398007645082822, + "grad_norm": 1.3991562128067017, + "learning_rate": 0.0001440158564680747, + "loss": 0.2942, + "step": 21750 + }, + { + "epoch": 0.840186879802309, + "grad_norm": 2.022538185119629, + "learning_rate": 0.00014399011544847292, + "loss": 0.2699, + "step": 21760 + }, + { + "epoch": 0.8405729950963358, + "grad_norm": 2.418179512023926, + "learning_rate": 0.00014396437442887113, + "loss": 0.3481, + "step": 21770 + }, + { + "epoch": 0.8409591103903625, + "grad_norm": 0.930482029914856, + "learning_rate": 0.00014393863340926935, + "loss": 0.3257, + "step": 21780 + }, + { + "epoch": 0.8413452256843894, + "grad_norm": 3.616676092147827, + "learning_rate": 0.00014391289238966756, + "loss": 0.3844, + "step": 21790 + }, + { + "epoch": 0.8417313409784162, + "grad_norm": 1.7993167638778687, + "learning_rate": 0.00014388715137006577, + "loss": 0.4569, + "step": 21800 + }, + { + "epoch": 0.8421174562724429, + "grad_norm": 1.9243824481964111, + "learning_rate": 0.000143861410350464, + "loss": 0.282, + "step": 21810 + }, + { + "epoch": 0.8425035715664697, + "grad_norm": 1.6578466892242432, + "learning_rate": 0.0001438356693308622, + "loss": 0.314, + "step": 21820 + }, + { + "epoch": 0.8428896868604966, + "grad_norm": 1.4833110570907593, + "learning_rate": 0.00014380992831126041, + "loss": 0.2698, + "step": 21830 + }, + { + "epoch": 0.8432758021545234, + "grad_norm": 1.9081813097000122, + "learning_rate": 0.00014378418729165863, + "loss": 0.2392, + "step": 21840 + }, + { + "epoch": 0.8436619174485501, + "grad_norm": 1.8436548709869385, + "learning_rate": 0.00014375844627205684, + "loss": 0.1906, + "step": 21850 + }, + { + "epoch": 0.8440480327425769, + "grad_norm": 4.679655075073242, + "learning_rate": 0.00014373270525245505, + "loss": 0.6446, + "step": 21860 + }, + { + "epoch": 0.8444341480366038, + "grad_norm": 1.8216800689697266, + "learning_rate": 0.0001437069642328533, + "loss": 0.3272, + "step": 21870 + }, + { + "epoch": 0.8448202633306305, + "grad_norm": 1.0107386112213135, + "learning_rate": 0.00014368122321325148, + "loss": 0.3003, + "step": 21880 + }, + { + "epoch": 0.8452063786246573, + "grad_norm": 0.9573041796684265, + "learning_rate": 0.0001436554821936497, + "loss": 0.1757, + "step": 21890 + }, + { + "epoch": 0.8455924939186841, + "grad_norm": 0.9367936253547668, + "learning_rate": 0.0001436297411740479, + "loss": 0.2166, + "step": 21900 + }, + { + "epoch": 0.845978609212711, + "grad_norm": 3.1247951984405518, + "learning_rate": 0.00014360400015444612, + "loss": 0.3488, + "step": 21910 + }, + { + "epoch": 0.8463647245067377, + "grad_norm": 3.9438281059265137, + "learning_rate": 0.00014357825913484433, + "loss": 0.4498, + "step": 21920 + }, + { + "epoch": 0.8467508398007645, + "grad_norm": 0.909572958946228, + "learning_rate": 0.00014355251811524255, + "loss": 0.2698, + "step": 21930 + }, + { + "epoch": 0.8471369550947913, + "grad_norm": 2.6619715690612793, + "learning_rate": 0.0001435267770956408, + "loss": 0.4204, + "step": 21940 + }, + { + "epoch": 0.8475230703888181, + "grad_norm": 0.6143421530723572, + "learning_rate": 0.00014350103607603897, + "loss": 0.3573, + "step": 21950 + }, + { + "epoch": 0.8479091856828449, + "grad_norm": 0.3222682476043701, + "learning_rate": 0.0001434752950564372, + "loss": 0.2172, + "step": 21960 + }, + { + "epoch": 0.8482953009768717, + "grad_norm": 1.772538185119629, + "learning_rate": 0.0001434495540368354, + "loss": 0.4203, + "step": 21970 + }, + { + "epoch": 0.8486814162708984, + "grad_norm": 1.6327133178710938, + "learning_rate": 0.0001434238130172336, + "loss": 0.2153, + "step": 21980 + }, + { + "epoch": 0.8490675315649253, + "grad_norm": 1.0445518493652344, + "learning_rate": 0.00014339807199763183, + "loss": 0.3392, + "step": 21990 + }, + { + "epoch": 0.8494536468589521, + "grad_norm": 3.6096575260162354, + "learning_rate": 0.00014337233097803004, + "loss": 0.2691, + "step": 22000 + }, + { + "epoch": 0.8498397621529789, + "grad_norm": 1.4343204498291016, + "learning_rate": 0.00014334658995842828, + "loss": 0.3118, + "step": 22010 + }, + { + "epoch": 0.8502258774470057, + "grad_norm": 1.0348806381225586, + "learning_rate": 0.00014332084893882647, + "loss": 0.2886, + "step": 22020 + }, + { + "epoch": 0.8506119927410325, + "grad_norm": 0.5164201855659485, + "learning_rate": 0.00014329510791922468, + "loss": 0.2943, + "step": 22030 + }, + { + "epoch": 0.8509981080350593, + "grad_norm": 1.8109897375106812, + "learning_rate": 0.0001432693668996229, + "loss": 0.4021, + "step": 22040 + }, + { + "epoch": 0.851384223329086, + "grad_norm": 2.7065579891204834, + "learning_rate": 0.0001432436258800211, + "loss": 0.371, + "step": 22050 + }, + { + "epoch": 0.8517703386231129, + "grad_norm": 2.3028764724731445, + "learning_rate": 0.00014321788486041935, + "loss": 0.5026, + "step": 22060 + }, + { + "epoch": 0.8521564539171397, + "grad_norm": 1.3945609331130981, + "learning_rate": 0.00014319214384081753, + "loss": 0.4444, + "step": 22070 + }, + { + "epoch": 0.8525425692111664, + "grad_norm": 2.407951593399048, + "learning_rate": 0.00014316640282121577, + "loss": 0.4465, + "step": 22080 + }, + { + "epoch": 0.8529286845051932, + "grad_norm": 4.120944976806641, + "learning_rate": 0.00014314066180161396, + "loss": 0.3142, + "step": 22090 + }, + { + "epoch": 0.8533147997992201, + "grad_norm": 1.8841919898986816, + "learning_rate": 0.00014311492078201217, + "loss": 0.3609, + "step": 22100 + }, + { + "epoch": 0.8537009150932469, + "grad_norm": 5.1519951820373535, + "learning_rate": 0.00014308917976241039, + "loss": 0.3062, + "step": 22110 + }, + { + "epoch": 0.8540870303872736, + "grad_norm": 2.7280924320220947, + "learning_rate": 0.0001430634387428086, + "loss": 0.3678, + "step": 22120 + }, + { + "epoch": 0.8544731456813004, + "grad_norm": 0.23237809538841248, + "learning_rate": 0.00014303769772320684, + "loss": 0.2979, + "step": 22130 + }, + { + "epoch": 0.8548592609753273, + "grad_norm": 1.0587934255599976, + "learning_rate": 0.00014301195670360503, + "loss": 0.5672, + "step": 22140 + }, + { + "epoch": 0.855245376269354, + "grad_norm": 1.854447603225708, + "learning_rate": 0.00014298621568400327, + "loss": 0.3657, + "step": 22150 + }, + { + "epoch": 0.8556314915633808, + "grad_norm": 0.9766449332237244, + "learning_rate": 0.00014296047466440145, + "loss": 0.3219, + "step": 22160 + }, + { + "epoch": 0.8560176068574076, + "grad_norm": 1.7281047105789185, + "learning_rate": 0.00014293473364479967, + "loss": 0.3485, + "step": 22170 + }, + { + "epoch": 0.8564037221514345, + "grad_norm": 1.8366886377334595, + "learning_rate": 0.0001429089926251979, + "loss": 0.2929, + "step": 22180 + }, + { + "epoch": 0.8567898374454612, + "grad_norm": 0.8708136677742004, + "learning_rate": 0.0001428832516055961, + "loss": 0.3633, + "step": 22190 + }, + { + "epoch": 0.857175952739488, + "grad_norm": 1.5010342597961426, + "learning_rate": 0.00014285751058599433, + "loss": 0.3646, + "step": 22200 + }, + { + "epoch": 0.8575620680335148, + "grad_norm": 1.7438324689865112, + "learning_rate": 0.00014283176956639252, + "loss": 0.3093, + "step": 22210 + }, + { + "epoch": 0.8579481833275416, + "grad_norm": 1.5954341888427734, + "learning_rate": 0.00014280602854679076, + "loss": 0.2435, + "step": 22220 + }, + { + "epoch": 0.8583342986215684, + "grad_norm": 2.5279555320739746, + "learning_rate": 0.00014278028752718895, + "loss": 0.2526, + "step": 22230 + }, + { + "epoch": 0.8587204139155952, + "grad_norm": 3.4773006439208984, + "learning_rate": 0.00014275454650758716, + "loss": 0.3763, + "step": 22240 + }, + { + "epoch": 0.8591065292096219, + "grad_norm": 0.25110548734664917, + "learning_rate": 0.0001427288054879854, + "loss": 0.2265, + "step": 22250 + }, + { + "epoch": 0.8594926445036488, + "grad_norm": 2.3060946464538574, + "learning_rate": 0.00014270306446838359, + "loss": 0.3756, + "step": 22260 + }, + { + "epoch": 0.8598787597976756, + "grad_norm": 2.206308364868164, + "learning_rate": 0.00014267732344878183, + "loss": 0.295, + "step": 22270 + }, + { + "epoch": 0.8602648750917024, + "grad_norm": 1.1059858798980713, + "learning_rate": 0.00014265158242918, + "loss": 0.2382, + "step": 22280 + }, + { + "epoch": 0.8606509903857292, + "grad_norm": 0.045407798141241074, + "learning_rate": 0.00014262584140957825, + "loss": 0.1725, + "step": 22290 + }, + { + "epoch": 0.861037105679756, + "grad_norm": 0.2532581686973572, + "learning_rate": 0.00014260010038997644, + "loss": 0.3089, + "step": 22300 + }, + { + "epoch": 0.8614232209737828, + "grad_norm": 0.8851459622383118, + "learning_rate": 0.00014257435937037468, + "loss": 0.2721, + "step": 22310 + }, + { + "epoch": 0.8618093362678095, + "grad_norm": 2.9988598823547363, + "learning_rate": 0.0001425486183507729, + "loss": 0.3854, + "step": 22320 + }, + { + "epoch": 0.8621954515618364, + "grad_norm": 1.888629674911499, + "learning_rate": 0.00014252287733117108, + "loss": 0.4472, + "step": 22330 + }, + { + "epoch": 0.8625815668558632, + "grad_norm": 0.9517232179641724, + "learning_rate": 0.00014249713631156932, + "loss": 0.3094, + "step": 22340 + }, + { + "epoch": 0.86296768214989, + "grad_norm": 1.4752097129821777, + "learning_rate": 0.0001424713952919675, + "loss": 0.2649, + "step": 22350 + }, + { + "epoch": 0.8633537974439167, + "grad_norm": 1.642285704612732, + "learning_rate": 0.00014244565427236575, + "loss": 0.4418, + "step": 22360 + }, + { + "epoch": 0.8637399127379436, + "grad_norm": 2.2177469730377197, + "learning_rate": 0.00014241991325276396, + "loss": 0.6865, + "step": 22370 + }, + { + "epoch": 0.8641260280319704, + "grad_norm": 0.9089158773422241, + "learning_rate": 0.00014239417223316217, + "loss": 0.1842, + "step": 22380 + }, + { + "epoch": 0.8645121433259971, + "grad_norm": 1.108091115951538, + "learning_rate": 0.00014236843121356039, + "loss": 0.3181, + "step": 22390 + }, + { + "epoch": 0.8648982586200239, + "grad_norm": 3.1317670345306396, + "learning_rate": 0.00014234269019395857, + "loss": 0.4798, + "step": 22400 + }, + { + "epoch": 0.8652843739140508, + "grad_norm": 1.0352108478546143, + "learning_rate": 0.0001423169491743568, + "loss": 0.3473, + "step": 22410 + }, + { + "epoch": 0.8656704892080775, + "grad_norm": 0.48546215891838074, + "learning_rate": 0.000142291208154755, + "loss": 0.2731, + "step": 22420 + }, + { + "epoch": 0.8660566045021043, + "grad_norm": 1.1608140468597412, + "learning_rate": 0.00014226546713515324, + "loss": 0.3202, + "step": 22430 + }, + { + "epoch": 0.8664427197961311, + "grad_norm": 0.19237665832042694, + "learning_rate": 0.00014223972611555145, + "loss": 0.4985, + "step": 22440 + }, + { + "epoch": 0.866828835090158, + "grad_norm": 0.12056539207696915, + "learning_rate": 0.00014221398509594967, + "loss": 0.2071, + "step": 22450 + }, + { + "epoch": 0.8672149503841847, + "grad_norm": 1.416548252105713, + "learning_rate": 0.00014218824407634788, + "loss": 0.2572, + "step": 22460 + }, + { + "epoch": 0.8676010656782115, + "grad_norm": 0.816148042678833, + "learning_rate": 0.00014216250305674607, + "loss": 0.2368, + "step": 22470 + }, + { + "epoch": 0.8679871809722383, + "grad_norm": 3.2394118309020996, + "learning_rate": 0.0001421367620371443, + "loss": 0.3768, + "step": 22480 + }, + { + "epoch": 0.8683732962662651, + "grad_norm": 0.7187336087226868, + "learning_rate": 0.00014211102101754252, + "loss": 0.3297, + "step": 22490 + }, + { + "epoch": 0.8687594115602919, + "grad_norm": 0.5154927372932434, + "learning_rate": 0.00014208527999794073, + "loss": 0.3301, + "step": 22500 + }, + { + "epoch": 0.8691455268543187, + "grad_norm": 1.0461368560791016, + "learning_rate": 0.00014205953897833895, + "loss": 0.335, + "step": 22510 + }, + { + "epoch": 0.8695316421483454, + "grad_norm": 0.9720495343208313, + "learning_rate": 0.00014203379795873716, + "loss": 0.2405, + "step": 22520 + }, + { + "epoch": 0.8699177574423723, + "grad_norm": 2.147216558456421, + "learning_rate": 0.00014200805693913537, + "loss": 0.3291, + "step": 22530 + }, + { + "epoch": 0.8703038727363991, + "grad_norm": 1.162614345550537, + "learning_rate": 0.00014198231591953356, + "loss": 0.2705, + "step": 22540 + }, + { + "epoch": 0.8706899880304259, + "grad_norm": 0.5020268559455872, + "learning_rate": 0.0001419565748999318, + "loss": 0.4403, + "step": 22550 + }, + { + "epoch": 0.8710761033244527, + "grad_norm": 3.198425531387329, + "learning_rate": 0.00014193083388033, + "loss": 0.2244, + "step": 22560 + }, + { + "epoch": 0.8714622186184795, + "grad_norm": 2.4578161239624023, + "learning_rate": 0.00014190509286072823, + "loss": 0.3993, + "step": 22570 + }, + { + "epoch": 0.8718483339125063, + "grad_norm": 1.4559924602508545, + "learning_rate": 0.00014187935184112644, + "loss": 0.2631, + "step": 22580 + }, + { + "epoch": 0.872234449206533, + "grad_norm": 0.6834856271743774, + "learning_rate": 0.00014185361082152465, + "loss": 0.3183, + "step": 22590 + }, + { + "epoch": 0.8726205645005599, + "grad_norm": 3.9718177318573, + "learning_rate": 0.00014182786980192287, + "loss": 0.439, + "step": 22600 + }, + { + "epoch": 0.8730066797945867, + "grad_norm": 1.7797685861587524, + "learning_rate": 0.00014180212878232105, + "loss": 0.2938, + "step": 22610 + }, + { + "epoch": 0.8733927950886134, + "grad_norm": 2.137479543685913, + "learning_rate": 0.0001417763877627193, + "loss": 0.4733, + "step": 22620 + }, + { + "epoch": 0.8737789103826402, + "grad_norm": 0.7577596306800842, + "learning_rate": 0.0001417506467431175, + "loss": 0.3599, + "step": 22630 + }, + { + "epoch": 0.8741650256766671, + "grad_norm": 0.8944536447525024, + "learning_rate": 0.00014172490572351572, + "loss": 0.3393, + "step": 22640 + }, + { + "epoch": 0.8745511409706939, + "grad_norm": 1.59170663356781, + "learning_rate": 0.00014169916470391393, + "loss": 0.3578, + "step": 22650 + }, + { + "epoch": 0.8749372562647206, + "grad_norm": 0.9852517247200012, + "learning_rate": 0.00014167342368431215, + "loss": 0.4203, + "step": 22660 + }, + { + "epoch": 0.8753233715587474, + "grad_norm": 1.0319880247116089, + "learning_rate": 0.00014164768266471036, + "loss": 0.3332, + "step": 22670 + }, + { + "epoch": 0.8757094868527743, + "grad_norm": 3.713357925415039, + "learning_rate": 0.00014162194164510857, + "loss": 0.412, + "step": 22680 + }, + { + "epoch": 0.876095602146801, + "grad_norm": 1.3197567462921143, + "learning_rate": 0.00014159620062550679, + "loss": 0.286, + "step": 22690 + }, + { + "epoch": 0.8764817174408278, + "grad_norm": 3.604928493499756, + "learning_rate": 0.000141570459605905, + "loss": 0.4206, + "step": 22700 + }, + { + "epoch": 0.8768678327348546, + "grad_norm": 3.1074795722961426, + "learning_rate": 0.0001415447185863032, + "loss": 0.2261, + "step": 22710 + }, + { + "epoch": 0.8772539480288815, + "grad_norm": 2.855581760406494, + "learning_rate": 0.00014151897756670143, + "loss": 0.6825, + "step": 22720 + }, + { + "epoch": 0.8776400633229082, + "grad_norm": 3.5756995677948, + "learning_rate": 0.00014149323654709964, + "loss": 0.4283, + "step": 22730 + }, + { + "epoch": 0.878026178616935, + "grad_norm": 1.4255709648132324, + "learning_rate": 0.00014146749552749785, + "loss": 0.3243, + "step": 22740 + }, + { + "epoch": 0.8784122939109618, + "grad_norm": 0.3095746338367462, + "learning_rate": 0.00014144175450789607, + "loss": 0.1052, + "step": 22750 + }, + { + "epoch": 0.8787984092049886, + "grad_norm": 1.4129611253738403, + "learning_rate": 0.00014141601348829428, + "loss": 0.2746, + "step": 22760 + }, + { + "epoch": 0.8791845244990154, + "grad_norm": 0.6448315382003784, + "learning_rate": 0.0001413902724686925, + "loss": 0.3312, + "step": 22770 + }, + { + "epoch": 0.8795706397930422, + "grad_norm": 1.6328849792480469, + "learning_rate": 0.0001413645314490907, + "loss": 0.2346, + "step": 22780 + }, + { + "epoch": 0.879956755087069, + "grad_norm": 1.0974128246307373, + "learning_rate": 0.00014133879042948892, + "loss": 0.2446, + "step": 22790 + }, + { + "epoch": 0.8803428703810958, + "grad_norm": 2.3657541275024414, + "learning_rate": 0.00014131304940988713, + "loss": 0.3439, + "step": 22800 + }, + { + "epoch": 0.8807289856751226, + "grad_norm": 0.8959445953369141, + "learning_rate": 0.00014128730839028535, + "loss": 0.2896, + "step": 22810 + }, + { + "epoch": 0.8811151009691494, + "grad_norm": 1.5202107429504395, + "learning_rate": 0.00014126156737068356, + "loss": 0.1951, + "step": 22820 + }, + { + "epoch": 0.8815012162631762, + "grad_norm": 1.3710687160491943, + "learning_rate": 0.00014123582635108177, + "loss": 0.3193, + "step": 22830 + }, + { + "epoch": 0.881887331557203, + "grad_norm": 2.18868088722229, + "learning_rate": 0.00014121008533147999, + "loss": 0.3535, + "step": 22840 + }, + { + "epoch": 0.8822734468512298, + "grad_norm": 0.8251023888587952, + "learning_rate": 0.0001411843443118782, + "loss": 0.2536, + "step": 22850 + }, + { + "epoch": 0.8826595621452565, + "grad_norm": 1.0674525499343872, + "learning_rate": 0.0001411586032922764, + "loss": 0.3482, + "step": 22860 + }, + { + "epoch": 0.8830456774392834, + "grad_norm": 2.1199145317077637, + "learning_rate": 0.00014113286227267463, + "loss": 0.201, + "step": 22870 + }, + { + "epoch": 0.8834317927333102, + "grad_norm": 1.1958723068237305, + "learning_rate": 0.00014110712125307284, + "loss": 0.2813, + "step": 22880 + }, + { + "epoch": 0.883817908027337, + "grad_norm": 1.7805982828140259, + "learning_rate": 0.00014108138023347105, + "loss": 0.5317, + "step": 22890 + }, + { + "epoch": 0.8842040233213637, + "grad_norm": 1.9648222923278809, + "learning_rate": 0.00014105563921386926, + "loss": 0.3022, + "step": 22900 + }, + { + "epoch": 0.8845901386153906, + "grad_norm": 0.9053369164466858, + "learning_rate": 0.00014102989819426748, + "loss": 0.2886, + "step": 22910 + }, + { + "epoch": 0.8849762539094174, + "grad_norm": 1.251861333847046, + "learning_rate": 0.0001410041571746657, + "loss": 0.2632, + "step": 22920 + }, + { + "epoch": 0.8853623692034441, + "grad_norm": 3.8411691188812256, + "learning_rate": 0.0001409784161550639, + "loss": 0.3056, + "step": 22930 + }, + { + "epoch": 0.8857484844974709, + "grad_norm": 0.5969072580337524, + "learning_rate": 0.00014095267513546212, + "loss": 0.3806, + "step": 22940 + }, + { + "epoch": 0.8861345997914978, + "grad_norm": 4.688140392303467, + "learning_rate": 0.00014092693411586033, + "loss": 0.3318, + "step": 22950 + }, + { + "epoch": 0.8865207150855245, + "grad_norm": 4.2694993019104, + "learning_rate": 0.00014090119309625854, + "loss": 0.34, + "step": 22960 + }, + { + "epoch": 0.8869068303795513, + "grad_norm": 2.4169955253601074, + "learning_rate": 0.00014087545207665676, + "loss": 0.466, + "step": 22970 + }, + { + "epoch": 0.8872929456735781, + "grad_norm": 0.044384077191352844, + "learning_rate": 0.00014084971105705497, + "loss": 0.3196, + "step": 22980 + }, + { + "epoch": 0.887679060967605, + "grad_norm": 0.6931707262992859, + "learning_rate": 0.0001408239700374532, + "loss": 0.2166, + "step": 22990 + }, + { + "epoch": 0.8880651762616317, + "grad_norm": 3.195596933364868, + "learning_rate": 0.0001407982290178514, + "loss": 0.404, + "step": 23000 + }, + { + "epoch": 0.8884512915556585, + "grad_norm": 2.055058002471924, + "learning_rate": 0.0001407724879982496, + "loss": 0.3685, + "step": 23010 + }, + { + "epoch": 0.8888374068496853, + "grad_norm": 0.7302665114402771, + "learning_rate": 0.00014074674697864782, + "loss": 0.2247, + "step": 23020 + }, + { + "epoch": 0.8892235221437121, + "grad_norm": 1.1183364391326904, + "learning_rate": 0.00014072100595904604, + "loss": 0.3699, + "step": 23030 + }, + { + "epoch": 0.8896096374377389, + "grad_norm": 1.040553092956543, + "learning_rate": 0.00014069526493944425, + "loss": 0.1791, + "step": 23040 + }, + { + "epoch": 0.8899957527317657, + "grad_norm": 1.5321402549743652, + "learning_rate": 0.00014066952391984246, + "loss": 0.2533, + "step": 23050 + }, + { + "epoch": 0.8903818680257926, + "grad_norm": 0.298433780670166, + "learning_rate": 0.0001406437829002407, + "loss": 0.2291, + "step": 23060 + }, + { + "epoch": 0.8907679833198193, + "grad_norm": 2.563689947128296, + "learning_rate": 0.0001406180418806389, + "loss": 0.4175, + "step": 23070 + }, + { + "epoch": 0.8911540986138461, + "grad_norm": 3.0614495277404785, + "learning_rate": 0.00014059230086103713, + "loss": 0.2875, + "step": 23080 + }, + { + "epoch": 0.8915402139078729, + "grad_norm": 0.4387970268726349, + "learning_rate": 0.00014056655984143532, + "loss": 0.2982, + "step": 23090 + }, + { + "epoch": 0.8919263292018997, + "grad_norm": 2.5590367317199707, + "learning_rate": 0.00014054081882183353, + "loss": 0.2858, + "step": 23100 + }, + { + "epoch": 0.8923124444959265, + "grad_norm": 0.6369298696517944, + "learning_rate": 0.00014051507780223174, + "loss": 0.2889, + "step": 23110 + }, + { + "epoch": 0.8926985597899533, + "grad_norm": 0.5210187435150146, + "learning_rate": 0.00014048933678262996, + "loss": 0.2442, + "step": 23120 + }, + { + "epoch": 0.89308467508398, + "grad_norm": 3.5073516368865967, + "learning_rate": 0.0001404635957630282, + "loss": 0.1733, + "step": 23130 + }, + { + "epoch": 0.8934707903780069, + "grad_norm": 2.1915247440338135, + "learning_rate": 0.00014043785474342638, + "loss": 0.4282, + "step": 23140 + }, + { + "epoch": 0.8938569056720337, + "grad_norm": 0.8182128071784973, + "learning_rate": 0.00014041211372382462, + "loss": 0.5501, + "step": 23150 + }, + { + "epoch": 0.8942430209660605, + "grad_norm": 0.4261817932128906, + "learning_rate": 0.0001403863727042228, + "loss": 0.2192, + "step": 23160 + }, + { + "epoch": 0.8946291362600872, + "grad_norm": 1.244523525238037, + "learning_rate": 0.00014036063168462102, + "loss": 0.2909, + "step": 23170 + }, + { + "epoch": 0.8950152515541141, + "grad_norm": 1.1659152507781982, + "learning_rate": 0.00014033489066501926, + "loss": 0.3119, + "step": 23180 + }, + { + "epoch": 0.8954013668481409, + "grad_norm": 0.4217310845851898, + "learning_rate": 0.00014030914964541745, + "loss": 0.2454, + "step": 23190 + }, + { + "epoch": 0.8957874821421676, + "grad_norm": 1.5685316324234009, + "learning_rate": 0.0001402834086258157, + "loss": 0.3609, + "step": 23200 + }, + { + "epoch": 0.8961735974361944, + "grad_norm": 2.6524040699005127, + "learning_rate": 0.00014025766760621388, + "loss": 0.2508, + "step": 23210 + }, + { + "epoch": 0.8965597127302213, + "grad_norm": 2.4932234287261963, + "learning_rate": 0.00014023192658661212, + "loss": 0.4047, + "step": 23220 + }, + { + "epoch": 0.896945828024248, + "grad_norm": 0.5093832015991211, + "learning_rate": 0.0001402061855670103, + "loss": 0.1639, + "step": 23230 + }, + { + "epoch": 0.8973319433182748, + "grad_norm": 1.1632994413375854, + "learning_rate": 0.00014018044454740852, + "loss": 0.2457, + "step": 23240 + }, + { + "epoch": 0.8977180586123016, + "grad_norm": 2.181727647781372, + "learning_rate": 0.00014015470352780676, + "loss": 0.3784, + "step": 23250 + }, + { + "epoch": 0.8981041739063285, + "grad_norm": 3.8110599517822266, + "learning_rate": 0.00014012896250820494, + "loss": 0.282, + "step": 23260 + }, + { + "epoch": 0.8984902892003552, + "grad_norm": 2.8994619846343994, + "learning_rate": 0.00014010322148860318, + "loss": 0.3801, + "step": 23270 + }, + { + "epoch": 0.898876404494382, + "grad_norm": 1.2624458074569702, + "learning_rate": 0.00014007748046900137, + "loss": 0.3718, + "step": 23280 + }, + { + "epoch": 0.8992625197884088, + "grad_norm": 1.5995053052902222, + "learning_rate": 0.0001400517394493996, + "loss": 0.1828, + "step": 23290 + }, + { + "epoch": 0.8996486350824356, + "grad_norm": 2.251941680908203, + "learning_rate": 0.0001400259984297978, + "loss": 0.4434, + "step": 23300 + }, + { + "epoch": 0.9000347503764624, + "grad_norm": 1.1319392919540405, + "learning_rate": 0.000140000257410196, + "loss": 0.2479, + "step": 23310 + }, + { + "epoch": 0.9004208656704892, + "grad_norm": 2.075227737426758, + "learning_rate": 0.00013997451639059425, + "loss": 0.3079, + "step": 23320 + }, + { + "epoch": 0.9008069809645161, + "grad_norm": 0.6504748463630676, + "learning_rate": 0.00013994877537099244, + "loss": 0.2899, + "step": 23330 + }, + { + "epoch": 0.9011930962585428, + "grad_norm": 2.1623177528381348, + "learning_rate": 0.00013992303435139068, + "loss": 0.2651, + "step": 23340 + }, + { + "epoch": 0.9015792115525696, + "grad_norm": 2.159290313720703, + "learning_rate": 0.00013989729333178886, + "loss": 0.2975, + "step": 23350 + }, + { + "epoch": 0.9019653268465964, + "grad_norm": 0.7650458216667175, + "learning_rate": 0.0001398715523121871, + "loss": 0.2699, + "step": 23360 + }, + { + "epoch": 0.9023514421406232, + "grad_norm": 4.838365077972412, + "learning_rate": 0.00013984581129258532, + "loss": 0.4633, + "step": 23370 + }, + { + "epoch": 0.90273755743465, + "grad_norm": 4.598055362701416, + "learning_rate": 0.0001398200702729835, + "loss": 0.5255, + "step": 23380 + }, + { + "epoch": 0.9031236727286768, + "grad_norm": 0.9883280396461487, + "learning_rate": 0.00013979432925338174, + "loss": 0.5096, + "step": 23390 + }, + { + "epoch": 0.9035097880227035, + "grad_norm": 2.1574087142944336, + "learning_rate": 0.00013976858823377993, + "loss": 0.2999, + "step": 23400 + }, + { + "epoch": 0.9038959033167304, + "grad_norm": 1.7071588039398193, + "learning_rate": 0.00013974284721417817, + "loss": 0.4066, + "step": 23410 + }, + { + "epoch": 0.9042820186107572, + "grad_norm": 1.5206272602081299, + "learning_rate": 0.00013971710619457636, + "loss": 0.2841, + "step": 23420 + }, + { + "epoch": 0.904668133904784, + "grad_norm": 0.8649633526802063, + "learning_rate": 0.0001396913651749746, + "loss": 0.2263, + "step": 23430 + }, + { + "epoch": 0.9050542491988107, + "grad_norm": 0.35130754113197327, + "learning_rate": 0.0001396656241553728, + "loss": 0.3575, + "step": 23440 + }, + { + "epoch": 0.9054403644928376, + "grad_norm": 0.6659330725669861, + "learning_rate": 0.000139639883135771, + "loss": 0.2895, + "step": 23450 + }, + { + "epoch": 0.9058264797868644, + "grad_norm": 1.1387370824813843, + "learning_rate": 0.00013961414211616924, + "loss": 0.3574, + "step": 23460 + }, + { + "epoch": 0.9062125950808911, + "grad_norm": 1.8786828517913818, + "learning_rate": 0.00013958840109656742, + "loss": 0.5127, + "step": 23470 + }, + { + "epoch": 0.9065987103749179, + "grad_norm": 1.1299179792404175, + "learning_rate": 0.00013956266007696566, + "loss": 0.2215, + "step": 23480 + }, + { + "epoch": 0.9069848256689448, + "grad_norm": 1.1256846189498901, + "learning_rate": 0.00013953691905736388, + "loss": 0.302, + "step": 23490 + }, + { + "epoch": 0.9073709409629716, + "grad_norm": 0.8697860836982727, + "learning_rate": 0.0001395111780377621, + "loss": 0.2846, + "step": 23500 + }, + { + "epoch": 0.9077570562569983, + "grad_norm": 1.4939324855804443, + "learning_rate": 0.0001394854370181603, + "loss": 0.2376, + "step": 23510 + }, + { + "epoch": 0.9081431715510251, + "grad_norm": 0.5483170747756958, + "learning_rate": 0.00013945969599855852, + "loss": 0.2546, + "step": 23520 + }, + { + "epoch": 0.908529286845052, + "grad_norm": 1.0681931972503662, + "learning_rate": 0.00013943395497895673, + "loss": 0.2236, + "step": 23530 + }, + { + "epoch": 0.9089154021390787, + "grad_norm": 1.9246234893798828, + "learning_rate": 0.00013940821395935492, + "loss": 0.3332, + "step": 23540 + }, + { + "epoch": 0.9093015174331055, + "grad_norm": 6.114970684051514, + "learning_rate": 0.00013938247293975316, + "loss": 0.3265, + "step": 23550 + }, + { + "epoch": 0.9096876327271323, + "grad_norm": 2.369112968444824, + "learning_rate": 0.00013935673192015137, + "loss": 0.3105, + "step": 23560 + }, + { + "epoch": 0.9100737480211591, + "grad_norm": 4.402872562408447, + "learning_rate": 0.00013933099090054958, + "loss": 0.3496, + "step": 23570 + }, + { + "epoch": 0.9104598633151859, + "grad_norm": 0.6064890027046204, + "learning_rate": 0.0001393052498809478, + "loss": 0.1322, + "step": 23580 + }, + { + "epoch": 0.9108459786092127, + "grad_norm": 0.41702714562416077, + "learning_rate": 0.000139279508861346, + "loss": 0.0886, + "step": 23590 + }, + { + "epoch": 0.9112320939032396, + "grad_norm": 1.1597472429275513, + "learning_rate": 0.00013925376784174422, + "loss": 0.1967, + "step": 23600 + }, + { + "epoch": 0.9116182091972663, + "grad_norm": 1.1049001216888428, + "learning_rate": 0.0001392280268221424, + "loss": 0.2802, + "step": 23610 + }, + { + "epoch": 0.9120043244912931, + "grad_norm": 0.7986807227134705, + "learning_rate": 0.00013920228580254065, + "loss": 0.1872, + "step": 23620 + }, + { + "epoch": 0.9123904397853199, + "grad_norm": 0.548693060874939, + "learning_rate": 0.00013917654478293886, + "loss": 0.3561, + "step": 23630 + }, + { + "epoch": 0.9127765550793467, + "grad_norm": 1.5944240093231201, + "learning_rate": 0.00013915080376333708, + "loss": 0.4619, + "step": 23640 + }, + { + "epoch": 0.9131626703733735, + "grad_norm": 1.9891632795333862, + "learning_rate": 0.0001391250627437353, + "loss": 0.3883, + "step": 23650 + }, + { + "epoch": 0.9135487856674003, + "grad_norm": 0.4564145803451538, + "learning_rate": 0.0001390993217241335, + "loss": 0.233, + "step": 23660 + }, + { + "epoch": 0.913934900961427, + "grad_norm": 1.1683684587478638, + "learning_rate": 0.00013907358070453172, + "loss": 0.4692, + "step": 23670 + }, + { + "epoch": 0.9143210162554539, + "grad_norm": 5.883500099182129, + "learning_rate": 0.00013904783968492993, + "loss": 0.2134, + "step": 23680 + }, + { + "epoch": 0.9147071315494807, + "grad_norm": 0.7426010370254517, + "learning_rate": 0.00013902209866532814, + "loss": 0.3608, + "step": 23690 + }, + { + "epoch": 0.9150932468435075, + "grad_norm": 1.5476068258285522, + "learning_rate": 0.00013899635764572636, + "loss": 0.2194, + "step": 23700 + }, + { + "epoch": 0.9154793621375342, + "grad_norm": 1.5702605247497559, + "learning_rate": 0.00013897061662612457, + "loss": 0.1731, + "step": 23710 + }, + { + "epoch": 0.9158654774315611, + "grad_norm": 2.336073637008667, + "learning_rate": 0.00013894487560652278, + "loss": 0.2703, + "step": 23720 + }, + { + "epoch": 0.9162515927255879, + "grad_norm": 0.4154629111289978, + "learning_rate": 0.000138919134586921, + "loss": 0.2601, + "step": 23730 + }, + { + "epoch": 0.9166377080196146, + "grad_norm": 1.9994091987609863, + "learning_rate": 0.0001388933935673192, + "loss": 0.4536, + "step": 23740 + }, + { + "epoch": 0.9170238233136414, + "grad_norm": 0.4610597491264343, + "learning_rate": 0.00013886765254771742, + "loss": 0.2726, + "step": 23750 + }, + { + "epoch": 0.9174099386076683, + "grad_norm": 2.19671893119812, + "learning_rate": 0.00013884191152811564, + "loss": 0.3418, + "step": 23760 + }, + { + "epoch": 0.917796053901695, + "grad_norm": 0.619023323059082, + "learning_rate": 0.00013881617050851385, + "loss": 0.2761, + "step": 23770 + }, + { + "epoch": 0.9181821691957218, + "grad_norm": 1.667083978652954, + "learning_rate": 0.00013879042948891206, + "loss": 0.2341, + "step": 23780 + }, + { + "epoch": 0.9185682844897486, + "grad_norm": 0.349020391702652, + "learning_rate": 0.00013876468846931028, + "loss": 0.244, + "step": 23790 + }, + { + "epoch": 0.9189543997837755, + "grad_norm": 3.2495415210723877, + "learning_rate": 0.0001387389474497085, + "loss": 0.2454, + "step": 23800 + }, + { + "epoch": 0.9193405150778022, + "grad_norm": 0.7900146842002869, + "learning_rate": 0.0001387132064301067, + "loss": 0.3209, + "step": 23810 + }, + { + "epoch": 0.919726630371829, + "grad_norm": 1.2435237169265747, + "learning_rate": 0.00013868746541050492, + "loss": 0.3719, + "step": 23820 + }, + { + "epoch": 0.9201127456658558, + "grad_norm": 0.7372536659240723, + "learning_rate": 0.00013866172439090313, + "loss": 0.2984, + "step": 23830 + }, + { + "epoch": 0.9204988609598826, + "grad_norm": 2.814180374145508, + "learning_rate": 0.00013863598337130134, + "loss": 0.5081, + "step": 23840 + }, + { + "epoch": 0.9208849762539094, + "grad_norm": 3.5411558151245117, + "learning_rate": 0.00013861024235169956, + "loss": 0.4475, + "step": 23850 + }, + { + "epoch": 0.9212710915479362, + "grad_norm": 0.41628485918045044, + "learning_rate": 0.00013858450133209777, + "loss": 0.1679, + "step": 23860 + }, + { + "epoch": 0.9216572068419631, + "grad_norm": 0.7951272130012512, + "learning_rate": 0.00013855876031249598, + "loss": 0.4346, + "step": 23870 + }, + { + "epoch": 0.9220433221359898, + "grad_norm": 0.6857497692108154, + "learning_rate": 0.0001385330192928942, + "loss": 0.2262, + "step": 23880 + }, + { + "epoch": 0.9224294374300166, + "grad_norm": 2.732487678527832, + "learning_rate": 0.0001385072782732924, + "loss": 0.3198, + "step": 23890 + }, + { + "epoch": 0.9228155527240434, + "grad_norm": 0.18741728365421295, + "learning_rate": 0.00013848153725369062, + "loss": 0.2409, + "step": 23900 + }, + { + "epoch": 0.9232016680180702, + "grad_norm": 0.46343281865119934, + "learning_rate": 0.00013845579623408884, + "loss": 0.2379, + "step": 23910 + }, + { + "epoch": 0.923587783312097, + "grad_norm": 1.7090940475463867, + "learning_rate": 0.00013843005521448705, + "loss": 0.2274, + "step": 23920 + }, + { + "epoch": 0.9239738986061238, + "grad_norm": 0.2000303715467453, + "learning_rate": 0.00013840431419488526, + "loss": 0.5171, + "step": 23930 + }, + { + "epoch": 0.9243600139001505, + "grad_norm": 1.6152868270874023, + "learning_rate": 0.00013837857317528348, + "loss": 0.1348, + "step": 23940 + }, + { + "epoch": 0.9247461291941774, + "grad_norm": 1.5346245765686035, + "learning_rate": 0.0001383528321556817, + "loss": 0.3708, + "step": 23950 + }, + { + "epoch": 0.9251322444882042, + "grad_norm": 2.1073787212371826, + "learning_rate": 0.0001383270911360799, + "loss": 0.3694, + "step": 23960 + }, + { + "epoch": 0.925518359782231, + "grad_norm": 12.8298921585083, + "learning_rate": 0.00013830135011647812, + "loss": 0.269, + "step": 23970 + }, + { + "epoch": 0.9259044750762577, + "grad_norm": 0.43689021468162537, + "learning_rate": 0.00013827560909687633, + "loss": 0.3099, + "step": 23980 + }, + { + "epoch": 0.9262905903702846, + "grad_norm": 2.084096908569336, + "learning_rate": 0.00013824986807727457, + "loss": 0.4423, + "step": 23990 + }, + { + "epoch": 0.9266767056643114, + "grad_norm": 0.9367966651916504, + "learning_rate": 0.00013822412705767276, + "loss": 0.2202, + "step": 24000 + }, + { + "epoch": 0.9270628209583381, + "grad_norm": 0.14286178350448608, + "learning_rate": 0.00013819838603807097, + "loss": 0.2302, + "step": 24010 + }, + { + "epoch": 0.9274489362523649, + "grad_norm": 0.7110779285430908, + "learning_rate": 0.00013817264501846918, + "loss": 0.3598, + "step": 24020 + }, + { + "epoch": 0.9278350515463918, + "grad_norm": 2.352980136871338, + "learning_rate": 0.0001381469039988674, + "loss": 0.4493, + "step": 24030 + }, + { + "epoch": 0.9282211668404186, + "grad_norm": 2.2235450744628906, + "learning_rate": 0.0001381211629792656, + "loss": 0.547, + "step": 24040 + }, + { + "epoch": 0.9286072821344453, + "grad_norm": 2.4419260025024414, + "learning_rate": 0.00013809542195966382, + "loss": 0.4612, + "step": 24050 + }, + { + "epoch": 0.9289933974284721, + "grad_norm": 1.3784935474395752, + "learning_rate": 0.00013806968094006206, + "loss": 0.3996, + "step": 24060 + }, + { + "epoch": 0.929379512722499, + "grad_norm": 1.019810676574707, + "learning_rate": 0.00013804393992046025, + "loss": 0.5654, + "step": 24070 + }, + { + "epoch": 0.9297656280165257, + "grad_norm": 2.399096965789795, + "learning_rate": 0.00013801819890085846, + "loss": 0.3335, + "step": 24080 + }, + { + "epoch": 0.9301517433105525, + "grad_norm": 0.28834161162376404, + "learning_rate": 0.00013799245788125668, + "loss": 0.4946, + "step": 24090 + }, + { + "epoch": 0.9305378586045794, + "grad_norm": 0.10426662117242813, + "learning_rate": 0.0001379667168616549, + "loss": 0.3275, + "step": 24100 + }, + { + "epoch": 0.9309239738986061, + "grad_norm": 2.7809340953826904, + "learning_rate": 0.0001379409758420531, + "loss": 0.3549, + "step": 24110 + }, + { + "epoch": 0.9313100891926329, + "grad_norm": 1.8708065748214722, + "learning_rate": 0.00013791523482245132, + "loss": 0.1429, + "step": 24120 + }, + { + "epoch": 0.9316962044866597, + "grad_norm": 1.9992274045944214, + "learning_rate": 0.00013788949380284956, + "loss": 0.2818, + "step": 24130 + }, + { + "epoch": 0.9320823197806866, + "grad_norm": 1.3665111064910889, + "learning_rate": 0.00013786375278324774, + "loss": 0.3565, + "step": 24140 + }, + { + "epoch": 0.9324684350747133, + "grad_norm": 3.106234550476074, + "learning_rate": 0.00013783801176364596, + "loss": 0.3706, + "step": 24150 + }, + { + "epoch": 0.9328545503687401, + "grad_norm": 1.2186559438705444, + "learning_rate": 0.00013781227074404417, + "loss": 0.1616, + "step": 24160 + }, + { + "epoch": 0.9332406656627669, + "grad_norm": 0.41551148891448975, + "learning_rate": 0.00013778652972444238, + "loss": 0.3073, + "step": 24170 + }, + { + "epoch": 0.9336267809567937, + "grad_norm": 2.015069007873535, + "learning_rate": 0.00013776078870484062, + "loss": 0.3317, + "step": 24180 + }, + { + "epoch": 0.9340128962508205, + "grad_norm": 1.9990328550338745, + "learning_rate": 0.0001377350476852388, + "loss": 0.4937, + "step": 24190 + }, + { + "epoch": 0.9343990115448473, + "grad_norm": 2.351898670196533, + "learning_rate": 0.00013770930666563705, + "loss": 0.3994, + "step": 24200 + }, + { + "epoch": 0.934785126838874, + "grad_norm": 1.4670008420944214, + "learning_rate": 0.00013768356564603524, + "loss": 0.2905, + "step": 24210 + }, + { + "epoch": 0.9351712421329009, + "grad_norm": 0.9890618920326233, + "learning_rate": 0.00013765782462643345, + "loss": 0.2512, + "step": 24220 + }, + { + "epoch": 0.9355573574269277, + "grad_norm": 0.3020402491092682, + "learning_rate": 0.00013763208360683166, + "loss": 0.3701, + "step": 24230 + }, + { + "epoch": 0.9359434727209545, + "grad_norm": 0.42218661308288574, + "learning_rate": 0.00013760634258722988, + "loss": 0.3395, + "step": 24240 + }, + { + "epoch": 0.9363295880149812, + "grad_norm": 1.5767306089401245, + "learning_rate": 0.00013758060156762812, + "loss": 0.3941, + "step": 24250 + }, + { + "epoch": 0.9367157033090081, + "grad_norm": 0.5416197180747986, + "learning_rate": 0.0001375548605480263, + "loss": 0.2915, + "step": 24260 + }, + { + "epoch": 0.9371018186030349, + "grad_norm": 1.175347924232483, + "learning_rate": 0.00013752911952842454, + "loss": 0.1284, + "step": 24270 + }, + { + "epoch": 0.9374879338970616, + "grad_norm": 0.8719255924224854, + "learning_rate": 0.00013750337850882273, + "loss": 0.3144, + "step": 24280 + }, + { + "epoch": 0.9378740491910884, + "grad_norm": 0.2711631655693054, + "learning_rate": 0.00013747763748922097, + "loss": 0.1561, + "step": 24290 + }, + { + "epoch": 0.9382601644851153, + "grad_norm": 3.2228004932403564, + "learning_rate": 0.00013745189646961918, + "loss": 0.5494, + "step": 24300 + }, + { + "epoch": 0.9386462797791421, + "grad_norm": 0.05647150054574013, + "learning_rate": 0.00013742615545001737, + "loss": 0.3571, + "step": 24310 + }, + { + "epoch": 0.9390323950731688, + "grad_norm": 0.8459005951881409, + "learning_rate": 0.0001374004144304156, + "loss": 0.1948, + "step": 24320 + }, + { + "epoch": 0.9394185103671956, + "grad_norm": 0.7044252157211304, + "learning_rate": 0.0001373746734108138, + "loss": 0.2511, + "step": 24330 + }, + { + "epoch": 0.9398046256612225, + "grad_norm": 0.979590654373169, + "learning_rate": 0.00013734893239121204, + "loss": 0.2158, + "step": 24340 + }, + { + "epoch": 0.9401907409552492, + "grad_norm": 1.0585628747940063, + "learning_rate": 0.00013732319137161022, + "loss": 0.1339, + "step": 24350 + }, + { + "epoch": 0.940576856249276, + "grad_norm": 0.6700488328933716, + "learning_rate": 0.00013729745035200846, + "loss": 0.2056, + "step": 24360 + }, + { + "epoch": 0.9409629715433029, + "grad_norm": 1.6656709909439087, + "learning_rate": 0.00013727170933240668, + "loss": 0.2957, + "step": 24370 + }, + { + "epoch": 0.9413490868373297, + "grad_norm": 2.963427782058716, + "learning_rate": 0.00013724596831280486, + "loss": 0.2673, + "step": 24380 + }, + { + "epoch": 0.9417352021313564, + "grad_norm": 0.8103615641593933, + "learning_rate": 0.0001372202272932031, + "loss": 0.372, + "step": 24390 + }, + { + "epoch": 0.9421213174253832, + "grad_norm": 2.3593873977661133, + "learning_rate": 0.0001371944862736013, + "loss": 0.431, + "step": 24400 + }, + { + "epoch": 0.9425074327194101, + "grad_norm": 2.8472931385040283, + "learning_rate": 0.00013716874525399953, + "loss": 0.4584, + "step": 24410 + }, + { + "epoch": 0.9428935480134368, + "grad_norm": 1.2765402793884277, + "learning_rate": 0.00013714300423439772, + "loss": 0.2543, + "step": 24420 + }, + { + "epoch": 0.9432796633074636, + "grad_norm": 1.4226797819137573, + "learning_rate": 0.00013711726321479596, + "loss": 0.2914, + "step": 24430 + }, + { + "epoch": 0.9436657786014904, + "grad_norm": 6.906572341918945, + "learning_rate": 0.00013709152219519417, + "loss": 0.4415, + "step": 24440 + }, + { + "epoch": 0.9440518938955172, + "grad_norm": 1.8387972116470337, + "learning_rate": 0.00013706578117559236, + "loss": 0.3018, + "step": 24450 + }, + { + "epoch": 0.944438009189544, + "grad_norm": 0.7259104251861572, + "learning_rate": 0.0001370400401559906, + "loss": 0.2356, + "step": 24460 + }, + { + "epoch": 0.9448241244835708, + "grad_norm": 0.6452949643135071, + "learning_rate": 0.00013701429913638878, + "loss": 0.2382, + "step": 24470 + }, + { + "epoch": 0.9452102397775975, + "grad_norm": 4.259208679199219, + "learning_rate": 0.00013698855811678702, + "loss": 0.4736, + "step": 24480 + }, + { + "epoch": 0.9455963550716244, + "grad_norm": 2.7305455207824707, + "learning_rate": 0.00013696281709718524, + "loss": 0.5034, + "step": 24490 + }, + { + "epoch": 0.9459824703656512, + "grad_norm": 0.6123724579811096, + "learning_rate": 0.00013693707607758345, + "loss": 0.3638, + "step": 24500 + }, + { + "epoch": 0.946368585659678, + "grad_norm": 0.9821889400482178, + "learning_rate": 0.00013691133505798166, + "loss": 0.3081, + "step": 24510 + }, + { + "epoch": 0.9467547009537047, + "grad_norm": 2.0870277881622314, + "learning_rate": 0.00013688559403837985, + "loss": 0.3116, + "step": 24520 + }, + { + "epoch": 0.9471408162477316, + "grad_norm": 2.495162010192871, + "learning_rate": 0.0001368598530187781, + "loss": 0.3879, + "step": 24530 + }, + { + "epoch": 0.9475269315417584, + "grad_norm": 1.4834142923355103, + "learning_rate": 0.00013683411199917628, + "loss": 0.1975, + "step": 24540 + }, + { + "epoch": 0.9479130468357851, + "grad_norm": 0.10831606388092041, + "learning_rate": 0.00013680837097957452, + "loss": 0.3504, + "step": 24550 + }, + { + "epoch": 0.9482991621298119, + "grad_norm": 1.6975635290145874, + "learning_rate": 0.00013678262995997273, + "loss": 0.3305, + "step": 24560 + }, + { + "epoch": 0.9486852774238388, + "grad_norm": 1.982422947883606, + "learning_rate": 0.00013675688894037094, + "loss": 0.3693, + "step": 24570 + }, + { + "epoch": 0.9490713927178656, + "grad_norm": 0.3487630784511566, + "learning_rate": 0.00013673114792076916, + "loss": 0.1912, + "step": 24580 + }, + { + "epoch": 0.9494575080118923, + "grad_norm": 0.4546245336532593, + "learning_rate": 0.00013670540690116734, + "loss": 0.2519, + "step": 24590 + }, + { + "epoch": 0.9498436233059191, + "grad_norm": 0.42953622341156006, + "learning_rate": 0.00013667966588156558, + "loss": 0.1967, + "step": 24600 + }, + { + "epoch": 0.950229738599946, + "grad_norm": 0.8197507858276367, + "learning_rate": 0.00013665392486196377, + "loss": 0.3057, + "step": 24610 + }, + { + "epoch": 0.9506158538939727, + "grad_norm": 0.8627083897590637, + "learning_rate": 0.000136628183842362, + "loss": 0.1382, + "step": 24620 + }, + { + "epoch": 0.9510019691879995, + "grad_norm": 1.0003200769424438, + "learning_rate": 0.00013660244282276022, + "loss": 0.1608, + "step": 24630 + }, + { + "epoch": 0.9513880844820264, + "grad_norm": 0.4473998546600342, + "learning_rate": 0.00013657670180315844, + "loss": 0.2226, + "step": 24640 + }, + { + "epoch": 0.9517741997760532, + "grad_norm": 1.9413338899612427, + "learning_rate": 0.00013655096078355665, + "loss": 0.2323, + "step": 24650 + }, + { + "epoch": 0.9521603150700799, + "grad_norm": 1.3250267505645752, + "learning_rate": 0.00013652521976395483, + "loss": 0.1617, + "step": 24660 + }, + { + "epoch": 0.9525464303641067, + "grad_norm": 1.756535530090332, + "learning_rate": 0.00013649947874435308, + "loss": 0.2727, + "step": 24670 + }, + { + "epoch": 0.9529325456581336, + "grad_norm": 1.1905356645584106, + "learning_rate": 0.0001364737377247513, + "loss": 0.3169, + "step": 24680 + }, + { + "epoch": 0.9533186609521603, + "grad_norm": 0.8116361498832703, + "learning_rate": 0.0001364479967051495, + "loss": 0.3999, + "step": 24690 + }, + { + "epoch": 0.9537047762461871, + "grad_norm": 0.8906353712081909, + "learning_rate": 0.00013642225568554772, + "loss": 0.3499, + "step": 24700 + }, + { + "epoch": 0.9540908915402139, + "grad_norm": 0.7422589659690857, + "learning_rate": 0.00013639651466594593, + "loss": 0.3544, + "step": 24710 + }, + { + "epoch": 0.9544770068342407, + "grad_norm": 0.21925519406795502, + "learning_rate": 0.00013637077364634414, + "loss": 0.1957, + "step": 24720 + }, + { + "epoch": 0.9548631221282675, + "grad_norm": 3.2993857860565186, + "learning_rate": 0.00013634503262674236, + "loss": 0.2625, + "step": 24730 + }, + { + "epoch": 0.9552492374222943, + "grad_norm": 1.4352943897247314, + "learning_rate": 0.00013631929160714057, + "loss": 0.1655, + "step": 24740 + }, + { + "epoch": 0.955635352716321, + "grad_norm": 1.4417182207107544, + "learning_rate": 0.00013629355058753878, + "loss": 0.2442, + "step": 24750 + }, + { + "epoch": 0.9560214680103479, + "grad_norm": 0.5024278163909912, + "learning_rate": 0.000136267809567937, + "loss": 0.2035, + "step": 24760 + }, + { + "epoch": 0.9564075833043747, + "grad_norm": 1.553472638130188, + "learning_rate": 0.0001362420685483352, + "loss": 0.22, + "step": 24770 + }, + { + "epoch": 0.9567936985984015, + "grad_norm": 1.655155897140503, + "learning_rate": 0.00013621632752873342, + "loss": 0.4051, + "step": 24780 + }, + { + "epoch": 0.9571798138924282, + "grad_norm": 0.4435586631298065, + "learning_rate": 0.00013619058650913164, + "loss": 0.3032, + "step": 24790 + }, + { + "epoch": 0.9575659291864551, + "grad_norm": 0.38589465618133545, + "learning_rate": 0.00013616484548952985, + "loss": 0.2062, + "step": 24800 + }, + { + "epoch": 0.9579520444804819, + "grad_norm": 0.6986583471298218, + "learning_rate": 0.00013613910446992806, + "loss": 0.6371, + "step": 24810 + }, + { + "epoch": 0.9583381597745086, + "grad_norm": 1.69257652759552, + "learning_rate": 0.00013611336345032627, + "loss": 0.2425, + "step": 24820 + }, + { + "epoch": 0.9587242750685354, + "grad_norm": 0.6136781573295593, + "learning_rate": 0.0001360876224307245, + "loss": 0.313, + "step": 24830 + }, + { + "epoch": 0.9591103903625623, + "grad_norm": 0.9019533395767212, + "learning_rate": 0.0001360618814111227, + "loss": 0.3723, + "step": 24840 + }, + { + "epoch": 0.9594965056565891, + "grad_norm": 4.601387977600098, + "learning_rate": 0.00013603614039152091, + "loss": 0.4228, + "step": 24850 + }, + { + "epoch": 0.9598826209506158, + "grad_norm": 0.35048994421958923, + "learning_rate": 0.00013601039937191913, + "loss": 0.2423, + "step": 24860 + }, + { + "epoch": 0.9602687362446426, + "grad_norm": 1.7177300453186035, + "learning_rate": 0.00013598465835231734, + "loss": 0.3995, + "step": 24870 + }, + { + "epoch": 0.9606548515386695, + "grad_norm": 1.3341178894042969, + "learning_rate": 0.00013595891733271555, + "loss": 0.3422, + "step": 24880 + }, + { + "epoch": 0.9610409668326962, + "grad_norm": 1.1859056949615479, + "learning_rate": 0.00013593317631311377, + "loss": 0.4204, + "step": 24890 + }, + { + "epoch": 0.961427082126723, + "grad_norm": 1.9831022024154663, + "learning_rate": 0.00013590743529351198, + "loss": 0.3623, + "step": 24900 + }, + { + "epoch": 0.9618131974207499, + "grad_norm": 0.8704162240028381, + "learning_rate": 0.0001358816942739102, + "loss": 0.2664, + "step": 24910 + }, + { + "epoch": 0.9621993127147767, + "grad_norm": 2.2464160919189453, + "learning_rate": 0.0001358559532543084, + "loss": 0.3628, + "step": 24920 + }, + { + "epoch": 0.9625854280088034, + "grad_norm": 2.5081570148468018, + "learning_rate": 0.00013583021223470662, + "loss": 0.314, + "step": 24930 + }, + { + "epoch": 0.9629715433028302, + "grad_norm": 4.44802713394165, + "learning_rate": 0.00013580447121510483, + "loss": 0.4593, + "step": 24940 + }, + { + "epoch": 0.9633576585968571, + "grad_norm": 2.0449907779693604, + "learning_rate": 0.00013577873019550305, + "loss": 0.3941, + "step": 24950 + }, + { + "epoch": 0.9637437738908838, + "grad_norm": 2.090315818786621, + "learning_rate": 0.00013575298917590126, + "loss": 0.3713, + "step": 24960 + }, + { + "epoch": 0.9641298891849106, + "grad_norm": 1.0789872407913208, + "learning_rate": 0.00013572724815629947, + "loss": 0.2411, + "step": 24970 + }, + { + "epoch": 0.9645160044789374, + "grad_norm": 2.0463109016418457, + "learning_rate": 0.0001357015071366977, + "loss": 0.2935, + "step": 24980 + }, + { + "epoch": 0.9649021197729643, + "grad_norm": 1.331299901008606, + "learning_rate": 0.00013567576611709593, + "loss": 0.2162, + "step": 24990 + }, + { + "epoch": 0.965288235066991, + "grad_norm": 3.3949038982391357, + "learning_rate": 0.00013565002509749411, + "loss": 0.4593, + "step": 25000 + }, + { + "epoch": 0.9656743503610178, + "grad_norm": 2.4510934352874756, + "learning_rate": 0.00013562428407789233, + "loss": 0.2476, + "step": 25010 + }, + { + "epoch": 0.9660604656550446, + "grad_norm": 1.1333917379379272, + "learning_rate": 0.00013559854305829054, + "loss": 0.3035, + "step": 25020 + }, + { + "epoch": 0.9664465809490714, + "grad_norm": 2.5525829792022705, + "learning_rate": 0.00013557280203868875, + "loss": 0.4498, + "step": 25030 + }, + { + "epoch": 0.9668326962430982, + "grad_norm": 1.4862838983535767, + "learning_rate": 0.00013554706101908697, + "loss": 0.3109, + "step": 25040 + }, + { + "epoch": 0.967218811537125, + "grad_norm": 1.0053727626800537, + "learning_rate": 0.00013552131999948518, + "loss": 0.3029, + "step": 25050 + }, + { + "epoch": 0.9676049268311517, + "grad_norm": 1.4033957719802856, + "learning_rate": 0.00013549557897988342, + "loss": 0.3096, + "step": 25060 + }, + { + "epoch": 0.9679910421251786, + "grad_norm": 2.1944756507873535, + "learning_rate": 0.0001354698379602816, + "loss": 0.3428, + "step": 25070 + }, + { + "epoch": 0.9683771574192054, + "grad_norm": 3.330143928527832, + "learning_rate": 0.00013544409694067982, + "loss": 0.4026, + "step": 25080 + }, + { + "epoch": 0.9687632727132321, + "grad_norm": 1.9337730407714844, + "learning_rate": 0.00013541835592107803, + "loss": 0.5343, + "step": 25090 + }, + { + "epoch": 0.9691493880072589, + "grad_norm": 4.203855514526367, + "learning_rate": 0.00013539261490147625, + "loss": 0.4461, + "step": 25100 + }, + { + "epoch": 0.9695355033012858, + "grad_norm": 0.6582885980606079, + "learning_rate": 0.0001353668738818745, + "loss": 0.2719, + "step": 25110 + }, + { + "epoch": 0.9699216185953126, + "grad_norm": 0.8788600564002991, + "learning_rate": 0.00013534113286227267, + "loss": 0.4335, + "step": 25120 + }, + { + "epoch": 0.9703077338893393, + "grad_norm": 1.3793160915374756, + "learning_rate": 0.00013531539184267091, + "loss": 0.3126, + "step": 25130 + }, + { + "epoch": 0.9706938491833662, + "grad_norm": 3.5996806621551514, + "learning_rate": 0.0001352896508230691, + "loss": 0.5171, + "step": 25140 + }, + { + "epoch": 0.971079964477393, + "grad_norm": 1.6220872402191162, + "learning_rate": 0.00013526390980346731, + "loss": 0.3992, + "step": 25150 + }, + { + "epoch": 0.9714660797714197, + "grad_norm": 1.8351634740829468, + "learning_rate": 0.00013523816878386553, + "loss": 0.3159, + "step": 25160 + }, + { + "epoch": 0.9718521950654465, + "grad_norm": 0.6400974988937378, + "learning_rate": 0.00013521242776426374, + "loss": 0.3187, + "step": 25170 + }, + { + "epoch": 0.9722383103594734, + "grad_norm": 1.3507485389709473, + "learning_rate": 0.00013518668674466198, + "loss": 0.3626, + "step": 25180 + }, + { + "epoch": 0.9726244256535002, + "grad_norm": 1.4778717756271362, + "learning_rate": 0.00013516094572506017, + "loss": 0.3467, + "step": 25190 + }, + { + "epoch": 0.9730105409475269, + "grad_norm": 0.4346179664134979, + "learning_rate": 0.0001351352047054584, + "loss": 0.1995, + "step": 25200 + }, + { + "epoch": 0.9733966562415537, + "grad_norm": 2.8404130935668945, + "learning_rate": 0.0001351094636858566, + "loss": 0.4765, + "step": 25210 + }, + { + "epoch": 0.9737827715355806, + "grad_norm": 0.044492240995168686, + "learning_rate": 0.0001350837226662548, + "loss": 0.3047, + "step": 25220 + }, + { + "epoch": 0.9741688868296073, + "grad_norm": 2.258355140686035, + "learning_rate": 0.00013505798164665302, + "loss": 0.234, + "step": 25230 + }, + { + "epoch": 0.9745550021236341, + "grad_norm": 0.23794110119342804, + "learning_rate": 0.00013503224062705123, + "loss": 0.5178, + "step": 25240 + }, + { + "epoch": 0.9749411174176609, + "grad_norm": 0.5849624872207642, + "learning_rate": 0.00013500649960744947, + "loss": 0.3419, + "step": 25250 + }, + { + "epoch": 0.9753272327116878, + "grad_norm": 1.535228967666626, + "learning_rate": 0.00013498075858784766, + "loss": 0.275, + "step": 25260 + }, + { + "epoch": 0.9757133480057145, + "grad_norm": 4.943759441375732, + "learning_rate": 0.0001349550175682459, + "loss": 0.241, + "step": 25270 + }, + { + "epoch": 0.9760994632997413, + "grad_norm": 1.3046916723251343, + "learning_rate": 0.0001349292765486441, + "loss": 0.2453, + "step": 25280 + }, + { + "epoch": 0.976485578593768, + "grad_norm": 0.20991156995296478, + "learning_rate": 0.0001349035355290423, + "loss": 0.392, + "step": 25290 + }, + { + "epoch": 0.9768716938877949, + "grad_norm": 1.2106267213821411, + "learning_rate": 0.00013487779450944054, + "loss": 0.284, + "step": 25300 + }, + { + "epoch": 0.9772578091818217, + "grad_norm": 0.28197771310806274, + "learning_rate": 0.00013485205348983873, + "loss": 0.2596, + "step": 25310 + }, + { + "epoch": 0.9776439244758485, + "grad_norm": 0.8538393378257751, + "learning_rate": 0.00013482631247023697, + "loss": 0.2519, + "step": 25320 + }, + { + "epoch": 0.9780300397698752, + "grad_norm": 1.9520586729049683, + "learning_rate": 0.00013480057145063515, + "loss": 0.2581, + "step": 25330 + }, + { + "epoch": 0.9784161550639021, + "grad_norm": 0.7613987326622009, + "learning_rate": 0.0001347748304310334, + "loss": 0.1549, + "step": 25340 + }, + { + "epoch": 0.9788022703579289, + "grad_norm": 0.2045626938343048, + "learning_rate": 0.00013474908941143158, + "loss": 0.2587, + "step": 25350 + }, + { + "epoch": 0.9791883856519557, + "grad_norm": 1.634488821029663, + "learning_rate": 0.0001347233483918298, + "loss": 0.2266, + "step": 25360 + }, + { + "epoch": 0.9795745009459824, + "grad_norm": 0.6475266814231873, + "learning_rate": 0.00013469760737222803, + "loss": 0.2548, + "step": 25370 + }, + { + "epoch": 0.9799606162400093, + "grad_norm": 0.25982800126075745, + "learning_rate": 0.00013467186635262622, + "loss": 0.3398, + "step": 25380 + }, + { + "epoch": 0.9803467315340361, + "grad_norm": 2.3229706287384033, + "learning_rate": 0.00013464612533302446, + "loss": 0.3468, + "step": 25390 + }, + { + "epoch": 0.9807328468280628, + "grad_norm": 0.6465128660202026, + "learning_rate": 0.00013462038431342265, + "loss": 0.2716, + "step": 25400 + }, + { + "epoch": 0.9811189621220897, + "grad_norm": 1.413368821144104, + "learning_rate": 0.0001345946432938209, + "loss": 0.2508, + "step": 25410 + }, + { + "epoch": 0.9815050774161165, + "grad_norm": 0.33577996492385864, + "learning_rate": 0.00013456890227421907, + "loss": 0.3323, + "step": 25420 + }, + { + "epoch": 0.9818911927101432, + "grad_norm": 0.7601230144500732, + "learning_rate": 0.0001345431612546173, + "loss": 0.2682, + "step": 25430 + }, + { + "epoch": 0.98227730800417, + "grad_norm": 4.296112060546875, + "learning_rate": 0.00013451742023501553, + "loss": 0.2886, + "step": 25440 + }, + { + "epoch": 0.9826634232981969, + "grad_norm": 1.2557302713394165, + "learning_rate": 0.00013449167921541371, + "loss": 0.2863, + "step": 25450 + }, + { + "epoch": 0.9830495385922237, + "grad_norm": 0.6168705821037292, + "learning_rate": 0.00013446593819581195, + "loss": 0.195, + "step": 25460 + }, + { + "epoch": 0.9834356538862504, + "grad_norm": 1.9064022302627563, + "learning_rate": 0.00013444019717621014, + "loss": 0.2898, + "step": 25470 + }, + { + "epoch": 0.9838217691802772, + "grad_norm": 0.3549353778362274, + "learning_rate": 0.00013441445615660838, + "loss": 0.1761, + "step": 25480 + }, + { + "epoch": 0.9842078844743041, + "grad_norm": 0.6308786869049072, + "learning_rate": 0.0001343887151370066, + "loss": 0.2169, + "step": 25490 + }, + { + "epoch": 0.9845939997683308, + "grad_norm": 0.9449920058250427, + "learning_rate": 0.0001343629741174048, + "loss": 0.2952, + "step": 25500 + }, + { + "epoch": 0.9849801150623576, + "grad_norm": 1.6993355751037598, + "learning_rate": 0.00013433723309780302, + "loss": 0.3745, + "step": 25510 + }, + { + "epoch": 0.9853662303563844, + "grad_norm": 0.8093920350074768, + "learning_rate": 0.0001343114920782012, + "loss": 0.1725, + "step": 25520 + }, + { + "epoch": 0.9857523456504113, + "grad_norm": 1.4968618154525757, + "learning_rate": 0.00013428575105859945, + "loss": 0.2843, + "step": 25530 + }, + { + "epoch": 0.986138460944438, + "grad_norm": 0.37341028451919556, + "learning_rate": 0.00013426001003899763, + "loss": 0.2462, + "step": 25540 + }, + { + "epoch": 0.9865245762384648, + "grad_norm": 2.0109541416168213, + "learning_rate": 0.00013423426901939587, + "loss": 0.4046, + "step": 25550 + }, + { + "epoch": 0.9869106915324916, + "grad_norm": 2.540151357650757, + "learning_rate": 0.0001342085279997941, + "loss": 0.4331, + "step": 25560 + }, + { + "epoch": 0.9872968068265184, + "grad_norm": 0.9178367257118225, + "learning_rate": 0.0001341827869801923, + "loss": 0.3849, + "step": 25570 + }, + { + "epoch": 0.9876829221205452, + "grad_norm": 1.4229514598846436, + "learning_rate": 0.00013415704596059051, + "loss": 0.326, + "step": 25580 + }, + { + "epoch": 0.988069037414572, + "grad_norm": 0.7699927091598511, + "learning_rate": 0.0001341313049409887, + "loss": 0.237, + "step": 25590 + }, + { + "epoch": 0.9884551527085987, + "grad_norm": 0.3460877239704132, + "learning_rate": 0.00013410556392138694, + "loss": 0.2679, + "step": 25600 + }, + { + "epoch": 0.9888412680026256, + "grad_norm": 0.25363796949386597, + "learning_rate": 0.00013407982290178515, + "loss": 0.1746, + "step": 25610 + }, + { + "epoch": 0.9892273832966524, + "grad_norm": 1.5607961416244507, + "learning_rate": 0.00013405408188218337, + "loss": 0.6328, + "step": 25620 + }, + { + "epoch": 0.9896134985906792, + "grad_norm": 2.3751626014709473, + "learning_rate": 0.00013402834086258158, + "loss": 0.3989, + "step": 25630 + }, + { + "epoch": 0.9899996138847059, + "grad_norm": 0.20423173904418945, + "learning_rate": 0.0001340025998429798, + "loss": 0.2836, + "step": 25640 + }, + { + "epoch": 0.9903857291787328, + "grad_norm": 1.4207524061203003, + "learning_rate": 0.000133976858823378, + "loss": 0.348, + "step": 25650 + }, + { + "epoch": 0.9907718444727596, + "grad_norm": 0.12217597663402557, + "learning_rate": 0.0001339511178037762, + "loss": 0.3515, + "step": 25660 + }, + { + "epoch": 0.9911579597667863, + "grad_norm": 0.9259626269340515, + "learning_rate": 0.00013392537678417443, + "loss": 0.3039, + "step": 25670 + }, + { + "epoch": 0.9915440750608132, + "grad_norm": 0.5700181722640991, + "learning_rate": 0.00013389963576457265, + "loss": 0.3266, + "step": 25680 + }, + { + "epoch": 0.99193019035484, + "grad_norm": 1.9731560945510864, + "learning_rate": 0.00013387389474497086, + "loss": 0.2534, + "step": 25690 + }, + { + "epoch": 0.9923163056488667, + "grad_norm": 1.3714967966079712, + "learning_rate": 0.00013384815372536907, + "loss": 0.298, + "step": 25700 + }, + { + "epoch": 0.9927024209428935, + "grad_norm": 1.2377171516418457, + "learning_rate": 0.0001338224127057673, + "loss": 0.2948, + "step": 25710 + }, + { + "epoch": 0.9930885362369204, + "grad_norm": 0.4711095690727234, + "learning_rate": 0.0001337966716861655, + "loss": 0.2295, + "step": 25720 + }, + { + "epoch": 0.9934746515309472, + "grad_norm": 0.46715909242630005, + "learning_rate": 0.00013377093066656369, + "loss": 0.3331, + "step": 25730 + }, + { + "epoch": 0.9938607668249739, + "grad_norm": 1.1080710887908936, + "learning_rate": 0.00013374518964696193, + "loss": 0.4324, + "step": 25740 + }, + { + "epoch": 0.9942468821190007, + "grad_norm": 2.3581650257110596, + "learning_rate": 0.00013371944862736014, + "loss": 0.372, + "step": 25750 + }, + { + "epoch": 0.9946329974130276, + "grad_norm": 2.448678970336914, + "learning_rate": 0.00013369370760775835, + "loss": 0.3396, + "step": 25760 + }, + { + "epoch": 0.9950191127070543, + "grad_norm": 1.4270198345184326, + "learning_rate": 0.00013366796658815657, + "loss": 0.4511, + "step": 25770 + }, + { + "epoch": 0.9954052280010811, + "grad_norm": 0.6360304951667786, + "learning_rate": 0.00013364222556855478, + "loss": 0.2427, + "step": 25780 + }, + { + "epoch": 0.9957913432951079, + "grad_norm": 2.1653332710266113, + "learning_rate": 0.000133616484548953, + "loss": 0.3057, + "step": 25790 + }, + { + "epoch": 0.9961774585891348, + "grad_norm": 2.952923536300659, + "learning_rate": 0.0001335907435293512, + "loss": 0.3076, + "step": 25800 + }, + { + "epoch": 0.9965635738831615, + "grad_norm": 0.5913527607917786, + "learning_rate": 0.00013356500250974942, + "loss": 0.399, + "step": 25810 + }, + { + "epoch": 0.9969496891771883, + "grad_norm": 0.4653400480747223, + "learning_rate": 0.00013353926149014763, + "loss": 0.2452, + "step": 25820 + }, + { + "epoch": 0.9973358044712151, + "grad_norm": 6.321722984313965, + "learning_rate": 0.00013351352047054585, + "loss": 0.3644, + "step": 25830 + }, + { + "epoch": 0.9977219197652419, + "grad_norm": 1.9944865703582764, + "learning_rate": 0.00013348777945094406, + "loss": 0.2915, + "step": 25840 + }, + { + "epoch": 0.9981080350592687, + "grad_norm": 1.1486843824386597, + "learning_rate": 0.00013346203843134227, + "loss": 0.3787, + "step": 25850 + }, + { + "epoch": 0.9984941503532955, + "grad_norm": 0.11532440781593323, + "learning_rate": 0.0001334362974117405, + "loss": 0.3053, + "step": 25860 + }, + { + "epoch": 0.9988802656473222, + "grad_norm": 1.3178479671478271, + "learning_rate": 0.0001334105563921387, + "loss": 0.2917, + "step": 25870 + }, + { + "epoch": 0.9992663809413491, + "grad_norm": 0.3511134088039398, + "learning_rate": 0.0001333848153725369, + "loss": 0.2657, + "step": 25880 + }, + { + "epoch": 0.9996524962353759, + "grad_norm": 1.2761729955673218, + "learning_rate": 0.00013335907435293513, + "loss": 0.3913, + "step": 25890 + }, + { + "epoch": 1.0000386115294027, + "grad_norm": 2.540947437286377, + "learning_rate": 0.00013333333333333334, + "loss": 0.3016, + "step": 25900 + }, + { + "epoch": 1.0004247268234294, + "grad_norm": 0.5865538120269775, + "learning_rate": 0.00013330759231373155, + "loss": 0.4184, + "step": 25910 + }, + { + "epoch": 1.0008108421174562, + "grad_norm": 2.7305166721343994, + "learning_rate": 0.00013328185129412977, + "loss": 0.3529, + "step": 25920 + }, + { + "epoch": 1.0011969574114832, + "grad_norm": 1.4203829765319824, + "learning_rate": 0.00013325611027452798, + "loss": 0.3345, + "step": 25930 + }, + { + "epoch": 1.00158307270551, + "grad_norm": 0.7687380313873291, + "learning_rate": 0.0001332303692549262, + "loss": 0.2622, + "step": 25940 + }, + { + "epoch": 1.0019691879995367, + "grad_norm": 0.4958217144012451, + "learning_rate": 0.0001332046282353244, + "loss": 0.3935, + "step": 25950 + }, + { + "epoch": 1.0023553032935635, + "grad_norm": 0.27102500200271606, + "learning_rate": 0.00013317888721572262, + "loss": 0.2642, + "step": 25960 + }, + { + "epoch": 1.0027414185875902, + "grad_norm": 0.6760912537574768, + "learning_rate": 0.00013315314619612083, + "loss": 0.1937, + "step": 25970 + }, + { + "epoch": 1.003127533881617, + "grad_norm": 2.5647270679473877, + "learning_rate": 0.00013312740517651905, + "loss": 0.3344, + "step": 25980 + }, + { + "epoch": 1.0035136491756438, + "grad_norm": 0.4810403883457184, + "learning_rate": 0.00013310166415691726, + "loss": 0.1783, + "step": 25990 + }, + { + "epoch": 1.0038997644696706, + "grad_norm": 2.5404248237609863, + "learning_rate": 0.00013307592313731547, + "loss": 0.3979, + "step": 26000 + }, + { + "epoch": 1.0042858797636975, + "grad_norm": 1.1135408878326416, + "learning_rate": 0.00013305018211771369, + "loss": 0.3681, + "step": 26010 + }, + { + "epoch": 1.0046719950577243, + "grad_norm": 1.2810723781585693, + "learning_rate": 0.0001330244410981119, + "loss": 0.183, + "step": 26020 + }, + { + "epoch": 1.005058110351751, + "grad_norm": 3.3486454486846924, + "learning_rate": 0.0001329987000785101, + "loss": 0.2489, + "step": 26030 + }, + { + "epoch": 1.0054442256457778, + "grad_norm": 0.7915325164794922, + "learning_rate": 0.00013297295905890833, + "loss": 0.3577, + "step": 26040 + }, + { + "epoch": 1.0058303409398046, + "grad_norm": 0.969727098941803, + "learning_rate": 0.00013294721803930654, + "loss": 0.2243, + "step": 26050 + }, + { + "epoch": 1.0062164562338314, + "grad_norm": 1.8932983875274658, + "learning_rate": 0.00013292147701970475, + "loss": 0.245, + "step": 26060 + }, + { + "epoch": 1.0066025715278581, + "grad_norm": 1.2421804666519165, + "learning_rate": 0.00013289573600010297, + "loss": 0.2087, + "step": 26070 + }, + { + "epoch": 1.006988686821885, + "grad_norm": 1.316405177116394, + "learning_rate": 0.00013286999498050118, + "loss": 0.2864, + "step": 26080 + }, + { + "epoch": 1.007374802115912, + "grad_norm": 1.9196691513061523, + "learning_rate": 0.0001328442539608994, + "loss": 0.2098, + "step": 26090 + }, + { + "epoch": 1.0077609174099387, + "grad_norm": 0.7522671222686768, + "learning_rate": 0.0001328185129412976, + "loss": 0.3092, + "step": 26100 + }, + { + "epoch": 1.0081470327039654, + "grad_norm": 0.322963684797287, + "learning_rate": 0.00013279277192169585, + "loss": 0.2452, + "step": 26110 + }, + { + "epoch": 1.0085331479979922, + "grad_norm": 2.243734121322632, + "learning_rate": 0.00013276703090209403, + "loss": 0.4387, + "step": 26120 + }, + { + "epoch": 1.008919263292019, + "grad_norm": 1.0588726997375488, + "learning_rate": 0.00013274128988249225, + "loss": 0.3326, + "step": 26130 + }, + { + "epoch": 1.0093053785860457, + "grad_norm": 0.7018478512763977, + "learning_rate": 0.00013271554886289046, + "loss": 0.4248, + "step": 26140 + }, + { + "epoch": 1.0096914938800725, + "grad_norm": 3.692521810531616, + "learning_rate": 0.00013268980784328867, + "loss": 0.2452, + "step": 26150 + }, + { + "epoch": 1.0100776091740993, + "grad_norm": 2.21842885017395, + "learning_rate": 0.00013266406682368689, + "loss": 0.3315, + "step": 26160 + }, + { + "epoch": 1.0104637244681263, + "grad_norm": 1.2321841716766357, + "learning_rate": 0.0001326383258040851, + "loss": 0.1042, + "step": 26170 + }, + { + "epoch": 1.010849839762153, + "grad_norm": 1.6151124238967896, + "learning_rate": 0.00013261258478448334, + "loss": 0.1934, + "step": 26180 + }, + { + "epoch": 1.0112359550561798, + "grad_norm": 0.0949881374835968, + "learning_rate": 0.00013258684376488153, + "loss": 0.1628, + "step": 26190 + }, + { + "epoch": 1.0116220703502066, + "grad_norm": 1.7337597608566284, + "learning_rate": 0.00013256110274527977, + "loss": 0.3275, + "step": 26200 + }, + { + "epoch": 1.0120081856442333, + "grad_norm": 2.1338372230529785, + "learning_rate": 0.00013253536172567795, + "loss": 0.3677, + "step": 26210 + }, + { + "epoch": 1.01239430093826, + "grad_norm": 1.803187370300293, + "learning_rate": 0.00013250962070607617, + "loss": 0.2886, + "step": 26220 + }, + { + "epoch": 1.0127804162322869, + "grad_norm": 2.022825002670288, + "learning_rate": 0.00013248387968647438, + "loss": 0.2618, + "step": 26230 + }, + { + "epoch": 1.0131665315263139, + "grad_norm": 0.48369279503822327, + "learning_rate": 0.0001324581386668726, + "loss": 0.4638, + "step": 26240 + }, + { + "epoch": 1.0135526468203406, + "grad_norm": 3.2530572414398193, + "learning_rate": 0.00013243239764727083, + "loss": 0.3824, + "step": 26250 + }, + { + "epoch": 1.0139387621143674, + "grad_norm": 1.0877609252929688, + "learning_rate": 0.00013240665662766902, + "loss": 0.1732, + "step": 26260 + }, + { + "epoch": 1.0143248774083942, + "grad_norm": 1.5585906505584717, + "learning_rate": 0.00013238091560806726, + "loss": 0.2044, + "step": 26270 + }, + { + "epoch": 1.014710992702421, + "grad_norm": 1.5326491594314575, + "learning_rate": 0.00013235517458846545, + "loss": 0.1759, + "step": 26280 + }, + { + "epoch": 1.0150971079964477, + "grad_norm": 4.376593589782715, + "learning_rate": 0.00013232943356886366, + "loss": 0.6032, + "step": 26290 + }, + { + "epoch": 1.0154832232904745, + "grad_norm": 0.8953253030776978, + "learning_rate": 0.0001323036925492619, + "loss": 0.4057, + "step": 26300 + }, + { + "epoch": 1.0158693385845012, + "grad_norm": 1.271932601928711, + "learning_rate": 0.00013227795152966009, + "loss": 0.1802, + "step": 26310 + }, + { + "epoch": 1.0162554538785282, + "grad_norm": 2.311713457107544, + "learning_rate": 0.00013225221051005833, + "loss": 0.3368, + "step": 26320 + }, + { + "epoch": 1.016641569172555, + "grad_norm": 1.386100172996521, + "learning_rate": 0.0001322264694904565, + "loss": 0.3101, + "step": 26330 + }, + { + "epoch": 1.0170276844665818, + "grad_norm": 1.371382236480713, + "learning_rate": 0.00013220072847085475, + "loss": 0.3804, + "step": 26340 + }, + { + "epoch": 1.0174137997606085, + "grad_norm": 0.7098391652107239, + "learning_rate": 0.00013217498745125294, + "loss": 0.2513, + "step": 26350 + }, + { + "epoch": 1.0177999150546353, + "grad_norm": 3.5874531269073486, + "learning_rate": 0.00013214924643165115, + "loss": 0.2467, + "step": 26360 + }, + { + "epoch": 1.018186030348662, + "grad_norm": 1.9853413105010986, + "learning_rate": 0.0001321235054120494, + "loss": 0.2255, + "step": 26370 + }, + { + "epoch": 1.0185721456426888, + "grad_norm": 0.5550156831741333, + "learning_rate": 0.00013209776439244758, + "loss": 0.1558, + "step": 26380 + }, + { + "epoch": 1.0189582609367156, + "grad_norm": 1.92972731590271, + "learning_rate": 0.00013207202337284582, + "loss": 0.4843, + "step": 26390 + }, + { + "epoch": 1.0193443762307426, + "grad_norm": 2.331674814224243, + "learning_rate": 0.000132046282353244, + "loss": 0.2652, + "step": 26400 + }, + { + "epoch": 1.0197304915247694, + "grad_norm": 0.809916615486145, + "learning_rate": 0.00013202054133364225, + "loss": 0.1824, + "step": 26410 + }, + { + "epoch": 1.0201166068187961, + "grad_norm": 2.6432926654815674, + "learning_rate": 0.00013199480031404046, + "loss": 0.2712, + "step": 26420 + }, + { + "epoch": 1.020502722112823, + "grad_norm": 1.6016955375671387, + "learning_rate": 0.00013196905929443865, + "loss": 0.1629, + "step": 26430 + }, + { + "epoch": 1.0208888374068497, + "grad_norm": 0.22881706058979034, + "learning_rate": 0.00013194331827483689, + "loss": 0.2977, + "step": 26440 + }, + { + "epoch": 1.0212749527008764, + "grad_norm": 1.3551994562149048, + "learning_rate": 0.00013191757725523507, + "loss": 0.1207, + "step": 26450 + }, + { + "epoch": 1.0216610679949032, + "grad_norm": 0.19062986969947815, + "learning_rate": 0.0001318918362356333, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 1.0220471832889302, + "grad_norm": 2.014047145843506, + "learning_rate": 0.0001318660952160315, + "loss": 0.4991, + "step": 26470 + }, + { + "epoch": 1.022433298582957, + "grad_norm": 0.39503228664398193, + "learning_rate": 0.00013184035419642974, + "loss": 0.3509, + "step": 26480 + }, + { + "epoch": 1.0228194138769837, + "grad_norm": 4.807271957397461, + "learning_rate": 0.00013181461317682795, + "loss": 0.3324, + "step": 26490 + }, + { + "epoch": 1.0232055291710105, + "grad_norm": 1.4785593748092651, + "learning_rate": 0.00013178887215722614, + "loss": 0.269, + "step": 26500 + }, + { + "epoch": 1.0235916444650373, + "grad_norm": 1.851137399673462, + "learning_rate": 0.00013176313113762438, + "loss": 0.2739, + "step": 26510 + }, + { + "epoch": 1.023977759759064, + "grad_norm": 0.6200979351997375, + "learning_rate": 0.00013173739011802257, + "loss": 0.3135, + "step": 26520 + }, + { + "epoch": 1.0243638750530908, + "grad_norm": 2.411592960357666, + "learning_rate": 0.0001317116490984208, + "loss": 0.1675, + "step": 26530 + }, + { + "epoch": 1.0247499903471176, + "grad_norm": 0.6965230703353882, + "learning_rate": 0.000131685908078819, + "loss": 0.2762, + "step": 26540 + }, + { + "epoch": 1.0251361056411445, + "grad_norm": 1.3620637655258179, + "learning_rate": 0.00013166016705921723, + "loss": 0.319, + "step": 26550 + }, + { + "epoch": 1.0255222209351713, + "grad_norm": 1.4016404151916504, + "learning_rate": 0.00013163442603961545, + "loss": 0.1871, + "step": 26560 + }, + { + "epoch": 1.025908336229198, + "grad_norm": 1.234718680381775, + "learning_rate": 0.00013160868502001363, + "loss": 0.1301, + "step": 26570 + }, + { + "epoch": 1.0262944515232248, + "grad_norm": 1.4657102823257446, + "learning_rate": 0.00013158294400041187, + "loss": 0.3814, + "step": 26580 + }, + { + "epoch": 1.0266805668172516, + "grad_norm": 1.606948733329773, + "learning_rate": 0.00013155720298081006, + "loss": 0.2226, + "step": 26590 + }, + { + "epoch": 1.0270666821112784, + "grad_norm": 0.4280283749103546, + "learning_rate": 0.0001315314619612083, + "loss": 0.1506, + "step": 26600 + }, + { + "epoch": 1.0274527974053052, + "grad_norm": 0.141262486577034, + "learning_rate": 0.0001315057209416065, + "loss": 0.1476, + "step": 26610 + }, + { + "epoch": 1.027838912699332, + "grad_norm": 2.7536983489990234, + "learning_rate": 0.00013147997992200473, + "loss": 0.2558, + "step": 26620 + }, + { + "epoch": 1.028225027993359, + "grad_norm": 1.052965760231018, + "learning_rate": 0.00013145423890240294, + "loss": 0.3989, + "step": 26630 + }, + { + "epoch": 1.0286111432873857, + "grad_norm": 0.3157159388065338, + "learning_rate": 0.00013142849788280112, + "loss": 0.2656, + "step": 26640 + }, + { + "epoch": 1.0289972585814124, + "grad_norm": 2.080801248550415, + "learning_rate": 0.00013140275686319937, + "loss": 0.3183, + "step": 26650 + }, + { + "epoch": 1.0293833738754392, + "grad_norm": 1.255540370941162, + "learning_rate": 0.00013137701584359755, + "loss": 0.4555, + "step": 26660 + }, + { + "epoch": 1.029769489169466, + "grad_norm": 1.0504742860794067, + "learning_rate": 0.0001313512748239958, + "loss": 0.17, + "step": 26670 + }, + { + "epoch": 1.0301556044634927, + "grad_norm": 1.8295503854751587, + "learning_rate": 0.000131325533804394, + "loss": 0.2202, + "step": 26680 + }, + { + "epoch": 1.0305417197575195, + "grad_norm": 0.9758415222167969, + "learning_rate": 0.00013129979278479222, + "loss": 0.1187, + "step": 26690 + }, + { + "epoch": 1.0309278350515463, + "grad_norm": 0.6076366901397705, + "learning_rate": 0.00013127405176519043, + "loss": 0.0668, + "step": 26700 + }, + { + "epoch": 1.0313139503455733, + "grad_norm": 0.7663784027099609, + "learning_rate": 0.00013124831074558865, + "loss": 0.2115, + "step": 26710 + }, + { + "epoch": 1.0317000656396, + "grad_norm": 1.814332365989685, + "learning_rate": 0.00013122256972598686, + "loss": 0.0901, + "step": 26720 + }, + { + "epoch": 1.0320861809336268, + "grad_norm": 2.80830454826355, + "learning_rate": 0.00013119682870638504, + "loss": 0.213, + "step": 26730 + }, + { + "epoch": 1.0324722962276536, + "grad_norm": 1.324601411819458, + "learning_rate": 0.00013117108768678328, + "loss": 0.6125, + "step": 26740 + }, + { + "epoch": 1.0328584115216803, + "grad_norm": 1.3301643133163452, + "learning_rate": 0.0001311453466671815, + "loss": 0.1986, + "step": 26750 + }, + { + "epoch": 1.033244526815707, + "grad_norm": 5.361929893493652, + "learning_rate": 0.0001311196056475797, + "loss": 0.5023, + "step": 26760 + }, + { + "epoch": 1.0336306421097339, + "grad_norm": 0.7855739593505859, + "learning_rate": 0.00013109386462797792, + "loss": 0.2549, + "step": 26770 + }, + { + "epoch": 1.0340167574037609, + "grad_norm": 0.05219104886054993, + "learning_rate": 0.00013106812360837614, + "loss": 0.2942, + "step": 26780 + }, + { + "epoch": 1.0344028726977876, + "grad_norm": 0.7680227160453796, + "learning_rate": 0.00013104238258877435, + "loss": 0.2909, + "step": 26790 + }, + { + "epoch": 1.0347889879918144, + "grad_norm": 0.559930682182312, + "learning_rate": 0.00013101664156917256, + "loss": 0.2876, + "step": 26800 + }, + { + "epoch": 1.0351751032858412, + "grad_norm": 0.8780495524406433, + "learning_rate": 0.00013099090054957078, + "loss": 0.2837, + "step": 26810 + }, + { + "epoch": 1.035561218579868, + "grad_norm": 2.0212693214416504, + "learning_rate": 0.000130965159529969, + "loss": 0.3008, + "step": 26820 + }, + { + "epoch": 1.0359473338738947, + "grad_norm": 2.2967641353607178, + "learning_rate": 0.0001309394185103672, + "loss": 0.2797, + "step": 26830 + }, + { + "epoch": 1.0363334491679215, + "grad_norm": 0.33352091908454895, + "learning_rate": 0.00013091367749076542, + "loss": 0.4785, + "step": 26840 + }, + { + "epoch": 1.0367195644619482, + "grad_norm": 1.5413645505905151, + "learning_rate": 0.00013088793647116363, + "loss": 0.3248, + "step": 26850 + }, + { + "epoch": 1.0371056797559752, + "grad_norm": 0.36845988035202026, + "learning_rate": 0.00013086219545156184, + "loss": 0.1723, + "step": 26860 + }, + { + "epoch": 1.037491795050002, + "grad_norm": 1.196103811264038, + "learning_rate": 0.00013083645443196006, + "loss": 0.3623, + "step": 26870 + }, + { + "epoch": 1.0378779103440288, + "grad_norm": 2.4172215461730957, + "learning_rate": 0.00013081071341235827, + "loss": 0.3428, + "step": 26880 + }, + { + "epoch": 1.0382640256380555, + "grad_norm": 1.585368275642395, + "learning_rate": 0.00013078497239275648, + "loss": 0.2791, + "step": 26890 + }, + { + "epoch": 1.0386501409320823, + "grad_norm": 2.8708138465881348, + "learning_rate": 0.0001307592313731547, + "loss": 0.3386, + "step": 26900 + }, + { + "epoch": 1.039036256226109, + "grad_norm": 1.1256765127182007, + "learning_rate": 0.0001307334903535529, + "loss": 0.2992, + "step": 26910 + }, + { + "epoch": 1.0394223715201358, + "grad_norm": 0.9846695065498352, + "learning_rate": 0.00013070774933395112, + "loss": 0.2659, + "step": 26920 + }, + { + "epoch": 1.0398084868141626, + "grad_norm": 1.5781525373458862, + "learning_rate": 0.00013068200831434934, + "loss": 0.2825, + "step": 26930 + }, + { + "epoch": 1.0401946021081896, + "grad_norm": 0.6615175008773804, + "learning_rate": 0.00013065626729474755, + "loss": 0.3608, + "step": 26940 + }, + { + "epoch": 1.0405807174022164, + "grad_norm": 1.2707170248031616, + "learning_rate": 0.00013063052627514576, + "loss": 0.3183, + "step": 26950 + }, + { + "epoch": 1.0409668326962431, + "grad_norm": 2.6396381855010986, + "learning_rate": 0.00013060478525554398, + "loss": 0.284, + "step": 26960 + }, + { + "epoch": 1.04135294799027, + "grad_norm": 1.5508041381835938, + "learning_rate": 0.0001305790442359422, + "loss": 0.2861, + "step": 26970 + }, + { + "epoch": 1.0417390632842967, + "grad_norm": 0.32720935344696045, + "learning_rate": 0.0001305533032163404, + "loss": 0.2784, + "step": 26980 + }, + { + "epoch": 1.0421251785783234, + "grad_norm": 2.3695240020751953, + "learning_rate": 0.00013052756219673862, + "loss": 0.3058, + "step": 26990 + }, + { + "epoch": 1.0425112938723502, + "grad_norm": 1.5404917001724243, + "learning_rate": 0.00013050182117713683, + "loss": 0.3636, + "step": 27000 + }, + { + "epoch": 1.0428974091663772, + "grad_norm": 1.3176417350769043, + "learning_rate": 0.00013047608015753504, + "loss": 0.2995, + "step": 27010 + }, + { + "epoch": 1.043283524460404, + "grad_norm": 1.5777894258499146, + "learning_rate": 0.00013045033913793326, + "loss": 0.3244, + "step": 27020 + }, + { + "epoch": 1.0436696397544307, + "grad_norm": 2.209652900695801, + "learning_rate": 0.00013042459811833147, + "loss": 0.2716, + "step": 27030 + }, + { + "epoch": 1.0440557550484575, + "grad_norm": 1.7991529703140259, + "learning_rate": 0.00013039885709872968, + "loss": 0.2022, + "step": 27040 + }, + { + "epoch": 1.0444418703424843, + "grad_norm": 0.3741607069969177, + "learning_rate": 0.0001303731160791279, + "loss": 0.3808, + "step": 27050 + }, + { + "epoch": 1.044827985636511, + "grad_norm": 0.023122821003198624, + "learning_rate": 0.0001303473750595261, + "loss": 0.3209, + "step": 27060 + }, + { + "epoch": 1.0452141009305378, + "grad_norm": 2.0373151302337646, + "learning_rate": 0.00013032163403992432, + "loss": 0.3105, + "step": 27070 + }, + { + "epoch": 1.0456002162245646, + "grad_norm": 0.5242247581481934, + "learning_rate": 0.00013029589302032254, + "loss": 0.2589, + "step": 27080 + }, + { + "epoch": 1.0459863315185915, + "grad_norm": 0.8598466515541077, + "learning_rate": 0.00013027015200072075, + "loss": 0.3025, + "step": 27090 + }, + { + "epoch": 1.0463724468126183, + "grad_norm": 0.19895502924919128, + "learning_rate": 0.00013024441098111896, + "loss": 0.3278, + "step": 27100 + }, + { + "epoch": 1.046758562106645, + "grad_norm": 1.9250338077545166, + "learning_rate": 0.0001302186699615172, + "loss": 0.3272, + "step": 27110 + }, + { + "epoch": 1.0471446774006719, + "grad_norm": 0.6927512288093567, + "learning_rate": 0.0001301929289419154, + "loss": 0.3256, + "step": 27120 + }, + { + "epoch": 1.0475307926946986, + "grad_norm": 0.5463778376579285, + "learning_rate": 0.0001301671879223136, + "loss": 0.2726, + "step": 27130 + }, + { + "epoch": 1.0479169079887254, + "grad_norm": 0.24690699577331543, + "learning_rate": 0.00013014144690271182, + "loss": 0.1247, + "step": 27140 + }, + { + "epoch": 1.0483030232827522, + "grad_norm": 7.875885009765625, + "learning_rate": 0.00013011570588311003, + "loss": 0.24, + "step": 27150 + }, + { + "epoch": 1.048689138576779, + "grad_norm": 2.0821642875671387, + "learning_rate": 0.00013008996486350824, + "loss": 0.2695, + "step": 27160 + }, + { + "epoch": 1.049075253870806, + "grad_norm": 1.0324410200119019, + "learning_rate": 0.00013006422384390646, + "loss": 0.3156, + "step": 27170 + }, + { + "epoch": 1.0494613691648327, + "grad_norm": 1.3675347566604614, + "learning_rate": 0.0001300384828243047, + "loss": 0.2366, + "step": 27180 + }, + { + "epoch": 1.0498474844588594, + "grad_norm": 0.4514729082584381, + "learning_rate": 0.00013001274180470288, + "loss": 0.1541, + "step": 27190 + }, + { + "epoch": 1.0502335997528862, + "grad_norm": 1.0098782777786255, + "learning_rate": 0.0001299870007851011, + "loss": 0.3059, + "step": 27200 + }, + { + "epoch": 1.050619715046913, + "grad_norm": 0.07944436371326447, + "learning_rate": 0.0001299612597654993, + "loss": 0.1882, + "step": 27210 + }, + { + "epoch": 1.0510058303409398, + "grad_norm": 0.8160178661346436, + "learning_rate": 0.00012993551874589752, + "loss": 0.2657, + "step": 27220 + }, + { + "epoch": 1.0513919456349665, + "grad_norm": 3.881469964981079, + "learning_rate": 0.00012990977772629574, + "loss": 0.4586, + "step": 27230 + }, + { + "epoch": 1.0517780609289935, + "grad_norm": 1.7851450443267822, + "learning_rate": 0.00012988403670669395, + "loss": 0.2945, + "step": 27240 + }, + { + "epoch": 1.0521641762230203, + "grad_norm": 4.176274299621582, + "learning_rate": 0.0001298582956870922, + "loss": 0.3819, + "step": 27250 + }, + { + "epoch": 1.052550291517047, + "grad_norm": 1.43479585647583, + "learning_rate": 0.00012983255466749038, + "loss": 0.3651, + "step": 27260 + }, + { + "epoch": 1.0529364068110738, + "grad_norm": 0.4839598536491394, + "learning_rate": 0.0001298068136478886, + "loss": 0.186, + "step": 27270 + }, + { + "epoch": 1.0533225221051006, + "grad_norm": 0.8487644791603088, + "learning_rate": 0.0001297810726282868, + "loss": 0.0851, + "step": 27280 + }, + { + "epoch": 1.0537086373991273, + "grad_norm": 3.5912392139434814, + "learning_rate": 0.00012975533160868502, + "loss": 0.3962, + "step": 27290 + }, + { + "epoch": 1.0540947526931541, + "grad_norm": 2.4018168449401855, + "learning_rate": 0.00012972959058908326, + "loss": 0.4, + "step": 27300 + }, + { + "epoch": 1.0544808679871809, + "grad_norm": 3.8987746238708496, + "learning_rate": 0.00012970384956948144, + "loss": 0.3081, + "step": 27310 + }, + { + "epoch": 1.0548669832812079, + "grad_norm": 0.4471427798271179, + "learning_rate": 0.00012967810854987968, + "loss": 0.2227, + "step": 27320 + }, + { + "epoch": 1.0552530985752346, + "grad_norm": 0.6207703351974487, + "learning_rate": 0.00012965236753027787, + "loss": 0.2066, + "step": 27330 + }, + { + "epoch": 1.0556392138692614, + "grad_norm": 1.4381637573242188, + "learning_rate": 0.00012962662651067608, + "loss": 0.2598, + "step": 27340 + }, + { + "epoch": 1.0560253291632882, + "grad_norm": 1.6057437658309937, + "learning_rate": 0.0001296008854910743, + "loss": 0.4003, + "step": 27350 + }, + { + "epoch": 1.056411444457315, + "grad_norm": 1.583121418952942, + "learning_rate": 0.0001295751444714725, + "loss": 0.2231, + "step": 27360 + }, + { + "epoch": 1.0567975597513417, + "grad_norm": 0.9260556101799011, + "learning_rate": 0.00012954940345187075, + "loss": 0.3066, + "step": 27370 + }, + { + "epoch": 1.0571836750453685, + "grad_norm": 0.530800461769104, + "learning_rate": 0.00012952366243226894, + "loss": 0.2063, + "step": 27380 + }, + { + "epoch": 1.0575697903393952, + "grad_norm": 0.3256929814815521, + "learning_rate": 0.00012949792141266718, + "loss": 0.2279, + "step": 27390 + }, + { + "epoch": 1.0579559056334222, + "grad_norm": 1.2628639936447144, + "learning_rate": 0.00012947218039306536, + "loss": 0.179, + "step": 27400 + }, + { + "epoch": 1.058342020927449, + "grad_norm": 1.168662190437317, + "learning_rate": 0.0001294464393734636, + "loss": 0.2862, + "step": 27410 + }, + { + "epoch": 1.0587281362214758, + "grad_norm": 3.8283517360687256, + "learning_rate": 0.00012942069835386182, + "loss": 0.3933, + "step": 27420 + }, + { + "epoch": 1.0591142515155025, + "grad_norm": 0.10550712049007416, + "learning_rate": 0.00012939495733426, + "loss": 0.2015, + "step": 27430 + }, + { + "epoch": 1.0595003668095293, + "grad_norm": 0.7177254557609558, + "learning_rate": 0.00012936921631465824, + "loss": 0.2126, + "step": 27440 + }, + { + "epoch": 1.059886482103556, + "grad_norm": 3.3265602588653564, + "learning_rate": 0.00012934347529505643, + "loss": 0.3734, + "step": 27450 + }, + { + "epoch": 1.0602725973975828, + "grad_norm": 0.9886051416397095, + "learning_rate": 0.00012931773427545467, + "loss": 0.2643, + "step": 27460 + }, + { + "epoch": 1.0606587126916098, + "grad_norm": 0.8335347175598145, + "learning_rate": 0.00012929199325585286, + "loss": 0.2044, + "step": 27470 + }, + { + "epoch": 1.0610448279856366, + "grad_norm": 1.16574227809906, + "learning_rate": 0.0001292662522362511, + "loss": 0.2316, + "step": 27480 + }, + { + "epoch": 1.0614309432796634, + "grad_norm": 0.1635606586933136, + "learning_rate": 0.0001292405112166493, + "loss": 0.2477, + "step": 27490 + }, + { + "epoch": 1.0618170585736901, + "grad_norm": 0.6800632476806641, + "learning_rate": 0.0001292147701970475, + "loss": 0.2831, + "step": 27500 + }, + { + "epoch": 1.062203173867717, + "grad_norm": 0.5231989622116089, + "learning_rate": 0.00012918902917744574, + "loss": 0.1621, + "step": 27510 + }, + { + "epoch": 1.0625892891617437, + "grad_norm": 6.668003559112549, + "learning_rate": 0.00012916328815784392, + "loss": 0.2736, + "step": 27520 + }, + { + "epoch": 1.0629754044557704, + "grad_norm": 0.2980963885784149, + "learning_rate": 0.00012913754713824216, + "loss": 0.2563, + "step": 27530 + }, + { + "epoch": 1.0633615197497972, + "grad_norm": 0.8144646883010864, + "learning_rate": 0.00012911180611864035, + "loss": 0.3821, + "step": 27540 + }, + { + "epoch": 1.063747635043824, + "grad_norm": 0.9781578183174133, + "learning_rate": 0.0001290860650990386, + "loss": 0.3653, + "step": 27550 + }, + { + "epoch": 1.064133750337851, + "grad_norm": 1.5652499198913574, + "learning_rate": 0.0001290603240794368, + "loss": 0.325, + "step": 27560 + }, + { + "epoch": 1.0645198656318777, + "grad_norm": 2.707165002822876, + "learning_rate": 0.000129034583059835, + "loss": 0.3129, + "step": 27570 + }, + { + "epoch": 1.0649059809259045, + "grad_norm": 0.35952532291412354, + "learning_rate": 0.00012900884204023323, + "loss": 0.2093, + "step": 27580 + }, + { + "epoch": 1.0652920962199313, + "grad_norm": 1.2863729000091553, + "learning_rate": 0.00012898310102063142, + "loss": 0.1299, + "step": 27590 + }, + { + "epoch": 1.065678211513958, + "grad_norm": 0.8279618620872498, + "learning_rate": 0.00012895736000102966, + "loss": 0.2072, + "step": 27600 + }, + { + "epoch": 1.0660643268079848, + "grad_norm": 0.4006168246269226, + "learning_rate": 0.00012893161898142787, + "loss": 0.3613, + "step": 27610 + }, + { + "epoch": 1.0664504421020116, + "grad_norm": 1.2714260816574097, + "learning_rate": 0.00012890587796182608, + "loss": 0.2013, + "step": 27620 + }, + { + "epoch": 1.0668365573960386, + "grad_norm": 1.9552396535873413, + "learning_rate": 0.0001288801369422243, + "loss": 0.2837, + "step": 27630 + }, + { + "epoch": 1.0672226726900653, + "grad_norm": 2.179871082305908, + "learning_rate": 0.00012885439592262248, + "loss": 0.1521, + "step": 27640 + }, + { + "epoch": 1.067608787984092, + "grad_norm": 1.1471878290176392, + "learning_rate": 0.00012882865490302072, + "loss": 0.2532, + "step": 27650 + }, + { + "epoch": 1.0679949032781189, + "grad_norm": 3.0032637119293213, + "learning_rate": 0.0001288029138834189, + "loss": 0.467, + "step": 27660 + }, + { + "epoch": 1.0683810185721456, + "grad_norm": 2.480180025100708, + "learning_rate": 0.00012877717286381715, + "loss": 0.3601, + "step": 27670 + }, + { + "epoch": 1.0687671338661724, + "grad_norm": 1.86027991771698, + "learning_rate": 0.00012875143184421536, + "loss": 0.3689, + "step": 27680 + }, + { + "epoch": 1.0691532491601992, + "grad_norm": 0.38396087288856506, + "learning_rate": 0.00012872569082461358, + "loss": 0.1401, + "step": 27690 + }, + { + "epoch": 1.0695393644542261, + "grad_norm": 0.8882033824920654, + "learning_rate": 0.0001286999498050118, + "loss": 0.6004, + "step": 27700 + }, + { + "epoch": 1.069925479748253, + "grad_norm": 1.2111278772354126, + "learning_rate": 0.00012867420878540998, + "loss": 0.2371, + "step": 27710 + }, + { + "epoch": 1.0703115950422797, + "grad_norm": 1.7789413928985596, + "learning_rate": 0.00012864846776580822, + "loss": 0.3142, + "step": 27720 + }, + { + "epoch": 1.0706977103363065, + "grad_norm": 1.6160372495651245, + "learning_rate": 0.00012862272674620643, + "loss": 0.3444, + "step": 27730 + }, + { + "epoch": 1.0710838256303332, + "grad_norm": 3.336289167404175, + "learning_rate": 0.00012859698572660464, + "loss": 0.1214, + "step": 27740 + }, + { + "epoch": 1.07146994092436, + "grad_norm": 4.104520320892334, + "learning_rate": 0.00012857124470700286, + "loss": 0.1752, + "step": 27750 + }, + { + "epoch": 1.0718560562183868, + "grad_norm": 0.4429762363433838, + "learning_rate": 0.00012854550368740107, + "loss": 0.5001, + "step": 27760 + }, + { + "epoch": 1.0722421715124135, + "grad_norm": 0.557033121585846, + "learning_rate": 0.00012851976266779928, + "loss": 0.3267, + "step": 27770 + }, + { + "epoch": 1.0726282868064403, + "grad_norm": 1.6847301721572876, + "learning_rate": 0.00012849402164819747, + "loss": 0.2305, + "step": 27780 + }, + { + "epoch": 1.0730144021004673, + "grad_norm": 1.0504320859909058, + "learning_rate": 0.0001284682806285957, + "loss": 0.3443, + "step": 27790 + }, + { + "epoch": 1.073400517394494, + "grad_norm": 2.728804349899292, + "learning_rate": 0.00012844253960899392, + "loss": 0.4083, + "step": 27800 + }, + { + "epoch": 1.0737866326885208, + "grad_norm": 0.6164497137069702, + "learning_rate": 0.00012841679858939214, + "loss": 0.2049, + "step": 27810 + }, + { + "epoch": 1.0741727479825476, + "grad_norm": 2.1831917762756348, + "learning_rate": 0.00012839105756979035, + "loss": 0.1674, + "step": 27820 + }, + { + "epoch": 1.0745588632765743, + "grad_norm": 0.5254467129707336, + "learning_rate": 0.00012836531655018856, + "loss": 0.3456, + "step": 27830 + }, + { + "epoch": 1.0749449785706011, + "grad_norm": 2.920846700668335, + "learning_rate": 0.00012833957553058678, + "loss": 0.4071, + "step": 27840 + }, + { + "epoch": 1.0753310938646279, + "grad_norm": 0.6006580591201782, + "learning_rate": 0.00012831383451098496, + "loss": 0.16, + "step": 27850 + }, + { + "epoch": 1.0757172091586549, + "grad_norm": 1.7163684368133545, + "learning_rate": 0.0001282880934913832, + "loss": 0.1821, + "step": 27860 + }, + { + "epoch": 1.0761033244526816, + "grad_norm": 1.8286449909210205, + "learning_rate": 0.00012826235247178142, + "loss": 0.3079, + "step": 27870 + }, + { + "epoch": 1.0764894397467084, + "grad_norm": 2.5178558826446533, + "learning_rate": 0.00012823661145217963, + "loss": 0.2158, + "step": 27880 + }, + { + "epoch": 1.0768755550407352, + "grad_norm": 0.8985245227813721, + "learning_rate": 0.00012821087043257784, + "loss": 0.198, + "step": 27890 + }, + { + "epoch": 1.077261670334762, + "grad_norm": 0.11768722534179688, + "learning_rate": 0.00012818512941297606, + "loss": 0.1661, + "step": 27900 + }, + { + "epoch": 1.0776477856287887, + "grad_norm": 1.0070226192474365, + "learning_rate": 0.00012815938839337427, + "loss": 0.2843, + "step": 27910 + }, + { + "epoch": 1.0780339009228155, + "grad_norm": 1.6243773698806763, + "learning_rate": 0.00012813364737377248, + "loss": 0.3449, + "step": 27920 + }, + { + "epoch": 1.0784200162168422, + "grad_norm": 2.8661181926727295, + "learning_rate": 0.0001281079063541707, + "loss": 0.4921, + "step": 27930 + }, + { + "epoch": 1.0788061315108692, + "grad_norm": 1.5015594959259033, + "learning_rate": 0.0001280821653345689, + "loss": 0.3347, + "step": 27940 + }, + { + "epoch": 1.079192246804896, + "grad_norm": 1.7244246006011963, + "learning_rate": 0.00012805642431496712, + "loss": 0.3342, + "step": 27950 + }, + { + "epoch": 1.0795783620989228, + "grad_norm": 0.17871785163879395, + "learning_rate": 0.00012803068329536534, + "loss": 0.051, + "step": 27960 + }, + { + "epoch": 1.0799644773929495, + "grad_norm": 1.093429684638977, + "learning_rate": 0.00012800494227576355, + "loss": 0.2461, + "step": 27970 + }, + { + "epoch": 1.0803505926869763, + "grad_norm": 1.0775126218795776, + "learning_rate": 0.00012797920125616176, + "loss": 0.293, + "step": 27980 + }, + { + "epoch": 1.080736707981003, + "grad_norm": 2.0808680057525635, + "learning_rate": 0.00012795346023655998, + "loss": 0.2125, + "step": 27990 + }, + { + "epoch": 1.0811228232750298, + "grad_norm": 3.172473907470703, + "learning_rate": 0.0001279277192169582, + "loss": 0.3289, + "step": 28000 + }, + { + "epoch": 1.0815089385690566, + "grad_norm": 0.8227205872535706, + "learning_rate": 0.0001279019781973564, + "loss": 0.4457, + "step": 28010 + }, + { + "epoch": 1.0818950538630836, + "grad_norm": 0.4987971782684326, + "learning_rate": 0.00012787623717775462, + "loss": 0.2397, + "step": 28020 + }, + { + "epoch": 1.0822811691571104, + "grad_norm": 0.6923367381095886, + "learning_rate": 0.00012785049615815283, + "loss": 0.1833, + "step": 28030 + }, + { + "epoch": 1.0826672844511371, + "grad_norm": 0.3719552457332611, + "learning_rate": 0.00012782475513855104, + "loss": 0.4045, + "step": 28040 + }, + { + "epoch": 1.083053399745164, + "grad_norm": 0.798744261264801, + "learning_rate": 0.00012779901411894926, + "loss": 0.2218, + "step": 28050 + }, + { + "epoch": 1.0834395150391907, + "grad_norm": 0.6289515495300293, + "learning_rate": 0.00012777327309934747, + "loss": 0.2423, + "step": 28060 + }, + { + "epoch": 1.0838256303332174, + "grad_norm": 0.6853532195091248, + "learning_rate": 0.00012774753207974568, + "loss": 0.2759, + "step": 28070 + }, + { + "epoch": 1.0842117456272442, + "grad_norm": 1.3380333185195923, + "learning_rate": 0.0001277217910601439, + "loss": 0.2457, + "step": 28080 + }, + { + "epoch": 1.0845978609212712, + "grad_norm": 1.4076060056686401, + "learning_rate": 0.0001276960500405421, + "loss": 0.3065, + "step": 28090 + }, + { + "epoch": 1.084983976215298, + "grad_norm": 1.790323257446289, + "learning_rate": 0.00012767030902094032, + "loss": 0.182, + "step": 28100 + }, + { + "epoch": 1.0853700915093247, + "grad_norm": 1.9291974306106567, + "learning_rate": 0.00012764456800133854, + "loss": 0.3798, + "step": 28110 + }, + { + "epoch": 1.0857562068033515, + "grad_norm": 1.36685049533844, + "learning_rate": 0.00012761882698173675, + "loss": 0.1893, + "step": 28120 + }, + { + "epoch": 1.0861423220973783, + "grad_norm": 2.52441668510437, + "learning_rate": 0.00012759308596213496, + "loss": 0.2389, + "step": 28130 + }, + { + "epoch": 1.086528437391405, + "grad_norm": 0.9578754901885986, + "learning_rate": 0.00012756734494253318, + "loss": 0.3133, + "step": 28140 + }, + { + "epoch": 1.0869145526854318, + "grad_norm": 0.17957572638988495, + "learning_rate": 0.0001275416039229314, + "loss": 0.1708, + "step": 28150 + }, + { + "epoch": 1.0873006679794586, + "grad_norm": 1.7213740348815918, + "learning_rate": 0.0001275158629033296, + "loss": 0.3831, + "step": 28160 + }, + { + "epoch": 1.0876867832734856, + "grad_norm": 1.2460767030715942, + "learning_rate": 0.00012749012188372782, + "loss": 0.266, + "step": 28170 + }, + { + "epoch": 1.0880728985675123, + "grad_norm": 0.33691835403442383, + "learning_rate": 0.00012746438086412606, + "loss": 0.216, + "step": 28180 + }, + { + "epoch": 1.088459013861539, + "grad_norm": 0.23887981474399567, + "learning_rate": 0.00012743863984452424, + "loss": 0.2331, + "step": 28190 + }, + { + "epoch": 1.0888451291555659, + "grad_norm": 4.753499984741211, + "learning_rate": 0.00012741289882492246, + "loss": 0.1845, + "step": 28200 + }, + { + "epoch": 1.0892312444495926, + "grad_norm": 1.055097222328186, + "learning_rate": 0.00012738715780532067, + "loss": 0.2978, + "step": 28210 + }, + { + "epoch": 1.0896173597436194, + "grad_norm": 2.3080852031707764, + "learning_rate": 0.00012736141678571888, + "loss": 0.1838, + "step": 28220 + }, + { + "epoch": 1.0900034750376462, + "grad_norm": 0.3733162581920624, + "learning_rate": 0.00012733567576611712, + "loss": 0.1941, + "step": 28230 + }, + { + "epoch": 1.090389590331673, + "grad_norm": 2.247748613357544, + "learning_rate": 0.0001273099347465153, + "loss": 0.2594, + "step": 28240 + }, + { + "epoch": 1.0907757056257, + "grad_norm": 1.644177794456482, + "learning_rate": 0.00012728419372691355, + "loss": 0.392, + "step": 28250 + }, + { + "epoch": 1.0911618209197267, + "grad_norm": 2.3522965908050537, + "learning_rate": 0.00012725845270731174, + "loss": 0.2198, + "step": 28260 + }, + { + "epoch": 1.0915479362137535, + "grad_norm": 0.2335210144519806, + "learning_rate": 0.00012723271168770995, + "loss": 0.3363, + "step": 28270 + }, + { + "epoch": 1.0919340515077802, + "grad_norm": 0.476607084274292, + "learning_rate": 0.00012720697066810816, + "loss": 0.1803, + "step": 28280 + }, + { + "epoch": 1.092320166801807, + "grad_norm": 2.1482882499694824, + "learning_rate": 0.00012718122964850638, + "loss": 0.2567, + "step": 28290 + }, + { + "epoch": 1.0927062820958338, + "grad_norm": 1.6457593441009521, + "learning_rate": 0.00012715548862890462, + "loss": 0.196, + "step": 28300 + }, + { + "epoch": 1.0930923973898605, + "grad_norm": 2.0742087364196777, + "learning_rate": 0.0001271297476093028, + "loss": 0.3955, + "step": 28310 + }, + { + "epoch": 1.0934785126838875, + "grad_norm": 0.1588711142539978, + "learning_rate": 0.00012710400658970104, + "loss": 0.2113, + "step": 28320 + }, + { + "epoch": 1.0938646279779143, + "grad_norm": 4.303687572479248, + "learning_rate": 0.00012707826557009923, + "loss": 0.2941, + "step": 28330 + }, + { + "epoch": 1.094250743271941, + "grad_norm": 2.0096209049224854, + "learning_rate": 0.00012705252455049744, + "loss": 0.3092, + "step": 28340 + }, + { + "epoch": 1.0946368585659678, + "grad_norm": 0.49071142077445984, + "learning_rate": 0.00012702678353089566, + "loss": 0.2387, + "step": 28350 + }, + { + "epoch": 1.0950229738599946, + "grad_norm": 0.9084739089012146, + "learning_rate": 0.00012700104251129387, + "loss": 0.2199, + "step": 28360 + }, + { + "epoch": 1.0954090891540214, + "grad_norm": 2.076706647872925, + "learning_rate": 0.0001269753014916921, + "loss": 0.3652, + "step": 28370 + }, + { + "epoch": 1.0957952044480481, + "grad_norm": 0.13036206364631653, + "learning_rate": 0.0001269495604720903, + "loss": 0.2514, + "step": 28380 + }, + { + "epoch": 1.0961813197420749, + "grad_norm": 0.146321102976799, + "learning_rate": 0.00012692381945248854, + "loss": 0.2721, + "step": 28390 + }, + { + "epoch": 1.0965674350361019, + "grad_norm": 0.8172006607055664, + "learning_rate": 0.00012689807843288672, + "loss": 0.16, + "step": 28400 + }, + { + "epoch": 1.0969535503301286, + "grad_norm": 1.099068522453308, + "learning_rate": 0.00012687233741328494, + "loss": 0.2489, + "step": 28410 + }, + { + "epoch": 1.0973396656241554, + "grad_norm": 0.6757088899612427, + "learning_rate": 0.00012684659639368318, + "loss": 0.1449, + "step": 28420 + }, + { + "epoch": 1.0977257809181822, + "grad_norm": 0.11124458909034729, + "learning_rate": 0.00012682085537408136, + "loss": 0.2202, + "step": 28430 + }, + { + "epoch": 1.098111896212209, + "grad_norm": 2.357466220855713, + "learning_rate": 0.0001267951143544796, + "loss": 0.2669, + "step": 28440 + }, + { + "epoch": 1.0984980115062357, + "grad_norm": 4.569977760314941, + "learning_rate": 0.0001267693733348778, + "loss": 0.3054, + "step": 28450 + }, + { + "epoch": 1.0988841268002625, + "grad_norm": 1.150667667388916, + "learning_rate": 0.00012674363231527603, + "loss": 0.1519, + "step": 28460 + }, + { + "epoch": 1.0992702420942893, + "grad_norm": 2.016101360321045, + "learning_rate": 0.00012671789129567421, + "loss": 0.2307, + "step": 28470 + }, + { + "epoch": 1.0996563573883162, + "grad_norm": 1.2213127613067627, + "learning_rate": 0.00012669215027607243, + "loss": 0.2847, + "step": 28480 + }, + { + "epoch": 1.100042472682343, + "grad_norm": 2.8080902099609375, + "learning_rate": 0.00012666640925647067, + "loss": 0.2295, + "step": 28490 + }, + { + "epoch": 1.1004285879763698, + "grad_norm": 1.4878045320510864, + "learning_rate": 0.00012664066823686885, + "loss": 0.1497, + "step": 28500 + }, + { + "epoch": 1.1008147032703965, + "grad_norm": 0.7453703880310059, + "learning_rate": 0.0001266149272172671, + "loss": 0.2052, + "step": 28510 + }, + { + "epoch": 1.1012008185644233, + "grad_norm": 0.2775499224662781, + "learning_rate": 0.00012658918619766528, + "loss": 0.1576, + "step": 28520 + }, + { + "epoch": 1.10158693385845, + "grad_norm": 1.0527644157409668, + "learning_rate": 0.00012656344517806352, + "loss": 0.2957, + "step": 28530 + }, + { + "epoch": 1.1019730491524768, + "grad_norm": 0.6511454582214355, + "learning_rate": 0.0001265377041584617, + "loss": 0.2336, + "step": 28540 + }, + { + "epoch": 1.1023591644465038, + "grad_norm": 0.32867324352264404, + "learning_rate": 0.00012651196313885992, + "loss": 0.1745, + "step": 28550 + }, + { + "epoch": 1.1027452797405306, + "grad_norm": 2.4408578872680664, + "learning_rate": 0.00012648622211925816, + "loss": 0.2195, + "step": 28560 + }, + { + "epoch": 1.1031313950345574, + "grad_norm": 1.4876518249511719, + "learning_rate": 0.00012646048109965635, + "loss": 0.5123, + "step": 28570 + }, + { + "epoch": 1.1035175103285841, + "grad_norm": 1.9403778314590454, + "learning_rate": 0.0001264347400800546, + "loss": 0.2783, + "step": 28580 + }, + { + "epoch": 1.103903625622611, + "grad_norm": 1.215280532836914, + "learning_rate": 0.00012640899906045277, + "loss": 0.2661, + "step": 28590 + }, + { + "epoch": 1.1042897409166377, + "grad_norm": 0.7179967164993286, + "learning_rate": 0.00012638325804085102, + "loss": 0.1756, + "step": 28600 + }, + { + "epoch": 1.1046758562106644, + "grad_norm": 1.9051718711853027, + "learning_rate": 0.00012635751702124923, + "loss": 0.2724, + "step": 28610 + }, + { + "epoch": 1.1050619715046912, + "grad_norm": 1.5659642219543457, + "learning_rate": 0.00012633177600164744, + "loss": 0.3891, + "step": 28620 + }, + { + "epoch": 1.1054480867987182, + "grad_norm": 0.9210501313209534, + "learning_rate": 0.00012630603498204566, + "loss": 0.1369, + "step": 28630 + }, + { + "epoch": 1.105834202092745, + "grad_norm": 0.5428475141525269, + "learning_rate": 0.00012628029396244384, + "loss": 0.1173, + "step": 28640 + }, + { + "epoch": 1.1062203173867717, + "grad_norm": 1.7399749755859375, + "learning_rate": 0.00012625455294284208, + "loss": 0.3005, + "step": 28650 + }, + { + "epoch": 1.1066064326807985, + "grad_norm": 0.09703828394412994, + "learning_rate": 0.00012622881192324027, + "loss": 0.3346, + "step": 28660 + }, + { + "epoch": 1.1069925479748253, + "grad_norm": 0.5834600329399109, + "learning_rate": 0.0001262030709036385, + "loss": 0.2689, + "step": 28670 + }, + { + "epoch": 1.107378663268852, + "grad_norm": 1.0677303075790405, + "learning_rate": 0.00012617732988403672, + "loss": 0.5609, + "step": 28680 + }, + { + "epoch": 1.1077647785628788, + "grad_norm": 1.6462419033050537, + "learning_rate": 0.00012615158886443493, + "loss": 0.1991, + "step": 28690 + }, + { + "epoch": 1.1081508938569056, + "grad_norm": 0.933779776096344, + "learning_rate": 0.00012612584784483315, + "loss": 0.2757, + "step": 28700 + }, + { + "epoch": 1.1085370091509326, + "grad_norm": 1.3413206338882446, + "learning_rate": 0.00012610010682523133, + "loss": 0.1798, + "step": 28710 + }, + { + "epoch": 1.1089231244449593, + "grad_norm": 4.479143142700195, + "learning_rate": 0.00012607436580562957, + "loss": 0.1777, + "step": 28720 + }, + { + "epoch": 1.109309239738986, + "grad_norm": 1.5768260955810547, + "learning_rate": 0.0001260486247860278, + "loss": 0.1967, + "step": 28730 + }, + { + "epoch": 1.1096953550330129, + "grad_norm": 2.1206741333007812, + "learning_rate": 0.000126022883766426, + "loss": 0.2399, + "step": 28740 + }, + { + "epoch": 1.1100814703270396, + "grad_norm": 1.4531667232513428, + "learning_rate": 0.00012599714274682421, + "loss": 0.3464, + "step": 28750 + }, + { + "epoch": 1.1104675856210664, + "grad_norm": 1.7988258600234985, + "learning_rate": 0.00012597140172722243, + "loss": 0.285, + "step": 28760 + }, + { + "epoch": 1.1108537009150932, + "grad_norm": 1.094808578491211, + "learning_rate": 0.00012594566070762064, + "loss": 0.2194, + "step": 28770 + }, + { + "epoch": 1.1112398162091202, + "grad_norm": 1.3884358406066895, + "learning_rate": 0.00012591991968801883, + "loss": 0.4267, + "step": 28780 + }, + { + "epoch": 1.111625931503147, + "grad_norm": 2.743480920791626, + "learning_rate": 0.00012589417866841707, + "loss": 0.3333, + "step": 28790 + }, + { + "epoch": 1.1120120467971737, + "grad_norm": 1.0373203754425049, + "learning_rate": 0.00012586843764881528, + "loss": 0.3941, + "step": 28800 + }, + { + "epoch": 1.1123981620912005, + "grad_norm": 2.018101692199707, + "learning_rate": 0.0001258426966292135, + "loss": 0.2928, + "step": 28810 + }, + { + "epoch": 1.1127842773852272, + "grad_norm": 2.567119836807251, + "learning_rate": 0.0001258169556096117, + "loss": 0.3597, + "step": 28820 + }, + { + "epoch": 1.113170392679254, + "grad_norm": 1.1235183477401733, + "learning_rate": 0.00012579121459000992, + "loss": 0.1807, + "step": 28830 + }, + { + "epoch": 1.1135565079732808, + "grad_norm": 1.3740451335906982, + "learning_rate": 0.00012576547357040813, + "loss": 0.2425, + "step": 28840 + }, + { + "epoch": 1.1139426232673075, + "grad_norm": 1.3751258850097656, + "learning_rate": 0.00012573973255080632, + "loss": 0.2116, + "step": 28850 + }, + { + "epoch": 1.1143287385613343, + "grad_norm": 0.9605401158332825, + "learning_rate": 0.00012571399153120456, + "loss": 0.2144, + "step": 28860 + }, + { + "epoch": 1.1147148538553613, + "grad_norm": 0.5127251148223877, + "learning_rate": 0.00012568825051160277, + "loss": 0.1902, + "step": 28870 + }, + { + "epoch": 1.115100969149388, + "grad_norm": 2.6720705032348633, + "learning_rate": 0.000125662509492001, + "loss": 0.3286, + "step": 28880 + }, + { + "epoch": 1.1154870844434148, + "grad_norm": 3.7196574211120605, + "learning_rate": 0.0001256367684723992, + "loss": 0.211, + "step": 28890 + }, + { + "epoch": 1.1158731997374416, + "grad_norm": 0.18796740472316742, + "learning_rate": 0.00012561102745279741, + "loss": 0.3351, + "step": 28900 + }, + { + "epoch": 1.1162593150314684, + "grad_norm": 1.3164410591125488, + "learning_rate": 0.00012558528643319563, + "loss": 0.1771, + "step": 28910 + }, + { + "epoch": 1.1166454303254951, + "grad_norm": 0.9552701115608215, + "learning_rate": 0.00012555954541359384, + "loss": 0.3302, + "step": 28920 + }, + { + "epoch": 1.117031545619522, + "grad_norm": 1.0072277784347534, + "learning_rate": 0.00012553380439399205, + "loss": 0.2936, + "step": 28930 + }, + { + "epoch": 1.1174176609135489, + "grad_norm": 0.24307872354984283, + "learning_rate": 0.00012550806337439027, + "loss": 0.2034, + "step": 28940 + }, + { + "epoch": 1.1178037762075757, + "grad_norm": 2.985166311264038, + "learning_rate": 0.00012548232235478848, + "loss": 0.2628, + "step": 28950 + }, + { + "epoch": 1.1181898915016024, + "grad_norm": 0.46555295586586, + "learning_rate": 0.0001254565813351867, + "loss": 0.1904, + "step": 28960 + }, + { + "epoch": 1.1185760067956292, + "grad_norm": 0.7813409566879272, + "learning_rate": 0.0001254308403155849, + "loss": 0.33, + "step": 28970 + }, + { + "epoch": 1.118962122089656, + "grad_norm": 2.4459455013275146, + "learning_rate": 0.00012540509929598312, + "loss": 0.3183, + "step": 28980 + }, + { + "epoch": 1.1193482373836827, + "grad_norm": 0.5164415240287781, + "learning_rate": 0.00012537935827638133, + "loss": 0.3959, + "step": 28990 + }, + { + "epoch": 1.1197343526777095, + "grad_norm": 0.3853105306625366, + "learning_rate": 0.00012535361725677955, + "loss": 0.1426, + "step": 29000 + }, + { + "epoch": 1.1201204679717365, + "grad_norm": 0.5817530751228333, + "learning_rate": 0.00012532787623717776, + "loss": 0.2218, + "step": 29010 + }, + { + "epoch": 1.1205065832657632, + "grad_norm": 1.264248251914978, + "learning_rate": 0.00012530213521757597, + "loss": 0.3002, + "step": 29020 + }, + { + "epoch": 1.12089269855979, + "grad_norm": 1.999251127243042, + "learning_rate": 0.0001252763941979742, + "loss": 0.3573, + "step": 29030 + }, + { + "epoch": 1.1212788138538168, + "grad_norm": 1.760797381401062, + "learning_rate": 0.0001252506531783724, + "loss": 0.3829, + "step": 29040 + }, + { + "epoch": 1.1216649291478435, + "grad_norm": 1.4757565259933472, + "learning_rate": 0.00012522491215877061, + "loss": 0.3407, + "step": 29050 + }, + { + "epoch": 1.1220510444418703, + "grad_norm": 0.08838029205799103, + "learning_rate": 0.00012519917113916883, + "loss": 0.1899, + "step": 29060 + }, + { + "epoch": 1.122437159735897, + "grad_norm": 0.6416037678718567, + "learning_rate": 0.00012517343011956704, + "loss": 0.277, + "step": 29070 + }, + { + "epoch": 1.1228232750299239, + "grad_norm": 2.9282822608947754, + "learning_rate": 0.00012514768909996525, + "loss": 0.3001, + "step": 29080 + }, + { + "epoch": 1.1232093903239506, + "grad_norm": 2.5984582901000977, + "learning_rate": 0.00012512194808036347, + "loss": 0.2326, + "step": 29090 + }, + { + "epoch": 1.1235955056179776, + "grad_norm": 1.0622142553329468, + "learning_rate": 0.00012509620706076168, + "loss": 0.1494, + "step": 29100 + }, + { + "epoch": 1.1239816209120044, + "grad_norm": 1.5386018753051758, + "learning_rate": 0.0001250704660411599, + "loss": 0.3392, + "step": 29110 + }, + { + "epoch": 1.1243677362060311, + "grad_norm": 0.8901385068893433, + "learning_rate": 0.0001250447250215581, + "loss": 0.2651, + "step": 29120 + }, + { + "epoch": 1.124753851500058, + "grad_norm": 2.0237483978271484, + "learning_rate": 0.00012501898400195632, + "loss": 0.3764, + "step": 29130 + }, + { + "epoch": 1.1251399667940847, + "grad_norm": 1.8989384174346924, + "learning_rate": 0.00012499324298235453, + "loss": 0.2713, + "step": 29140 + }, + { + "epoch": 1.1255260820881114, + "grad_norm": 2.704643487930298, + "learning_rate": 0.00012496750196275275, + "loss": 0.1362, + "step": 29150 + }, + { + "epoch": 1.1259121973821382, + "grad_norm": 1.2598273754119873, + "learning_rate": 0.00012494176094315096, + "loss": 0.331, + "step": 29160 + }, + { + "epoch": 1.1262983126761652, + "grad_norm": 2.2073826789855957, + "learning_rate": 0.00012491601992354917, + "loss": 0.209, + "step": 29170 + }, + { + "epoch": 1.126684427970192, + "grad_norm": 0.8338522911071777, + "learning_rate": 0.0001248902789039474, + "loss": 0.2583, + "step": 29180 + }, + { + "epoch": 1.1270705432642187, + "grad_norm": 0.49807825684547424, + "learning_rate": 0.0001248645378843456, + "loss": 0.2702, + "step": 29190 + }, + { + "epoch": 1.1274566585582455, + "grad_norm": 2.2561802864074707, + "learning_rate": 0.00012483879686474381, + "loss": 0.2667, + "step": 29200 + }, + { + "epoch": 1.1278427738522723, + "grad_norm": 0.6450731754302979, + "learning_rate": 0.00012481305584514203, + "loss": 0.2592, + "step": 29210 + }, + { + "epoch": 1.128228889146299, + "grad_norm": 1.214436650276184, + "learning_rate": 0.00012478731482554024, + "loss": 0.2285, + "step": 29220 + }, + { + "epoch": 1.1286150044403258, + "grad_norm": 1.2677173614501953, + "learning_rate": 0.00012476157380593848, + "loss": 0.1732, + "step": 29230 + }, + { + "epoch": 1.1290011197343528, + "grad_norm": 2.034266471862793, + "learning_rate": 0.00012473583278633667, + "loss": 0.3679, + "step": 29240 + }, + { + "epoch": 1.1293872350283796, + "grad_norm": 2.005913496017456, + "learning_rate": 0.00012471009176673488, + "loss": 0.2703, + "step": 29250 + }, + { + "epoch": 1.1297733503224063, + "grad_norm": 1.818928599357605, + "learning_rate": 0.0001246843507471331, + "loss": 0.3355, + "step": 29260 + }, + { + "epoch": 1.130159465616433, + "grad_norm": 0.5393241047859192, + "learning_rate": 0.0001246586097275313, + "loss": 0.1043, + "step": 29270 + }, + { + "epoch": 1.1305455809104599, + "grad_norm": 0.5508402585983276, + "learning_rate": 0.00012463286870792952, + "loss": 0.2538, + "step": 29280 + }, + { + "epoch": 1.1309316962044866, + "grad_norm": 1.1734035015106201, + "learning_rate": 0.00012460712768832773, + "loss": 0.3082, + "step": 29290 + }, + { + "epoch": 1.1313178114985134, + "grad_norm": 0.977611243724823, + "learning_rate": 0.00012458138666872597, + "loss": 0.1962, + "step": 29300 + }, + { + "epoch": 1.1317039267925402, + "grad_norm": 0.9720492362976074, + "learning_rate": 0.00012455564564912416, + "loss": 0.3462, + "step": 29310 + }, + { + "epoch": 1.132090042086567, + "grad_norm": 1.20888352394104, + "learning_rate": 0.00012452990462952237, + "loss": 0.1229, + "step": 29320 + }, + { + "epoch": 1.132476157380594, + "grad_norm": 0.7969954609870911, + "learning_rate": 0.0001245041636099206, + "loss": 0.4213, + "step": 29330 + }, + { + "epoch": 1.1328622726746207, + "grad_norm": 0.07595942914485931, + "learning_rate": 0.0001244784225903188, + "loss": 0.2148, + "step": 29340 + }, + { + "epoch": 1.1332483879686475, + "grad_norm": 0.15456156432628632, + "learning_rate": 0.000124452681570717, + "loss": 0.4713, + "step": 29350 + }, + { + "epoch": 1.1336345032626742, + "grad_norm": 1.232366681098938, + "learning_rate": 0.00012442694055111523, + "loss": 0.2446, + "step": 29360 + }, + { + "epoch": 1.134020618556701, + "grad_norm": 2.3669209480285645, + "learning_rate": 0.00012440119953151347, + "loss": 0.3025, + "step": 29370 + }, + { + "epoch": 1.1344067338507278, + "grad_norm": 4.639179229736328, + "learning_rate": 0.00012437545851191165, + "loss": 0.337, + "step": 29380 + }, + { + "epoch": 1.1347928491447545, + "grad_norm": 0.700533926486969, + "learning_rate": 0.0001243497174923099, + "loss": 0.1747, + "step": 29390 + }, + { + "epoch": 1.1351789644387815, + "grad_norm": 0.5738794803619385, + "learning_rate": 0.00012432397647270808, + "loss": 0.357, + "step": 29400 + }, + { + "epoch": 1.1355650797328083, + "grad_norm": 2.620095729827881, + "learning_rate": 0.0001242982354531063, + "loss": 0.2885, + "step": 29410 + }, + { + "epoch": 1.135951195026835, + "grad_norm": 1.5040203332901, + "learning_rate": 0.00012427249443350453, + "loss": 0.2481, + "step": 29420 + }, + { + "epoch": 1.1363373103208618, + "grad_norm": 0.7409051060676575, + "learning_rate": 0.00012424675341390272, + "loss": 0.3365, + "step": 29430 + }, + { + "epoch": 1.1367234256148886, + "grad_norm": 0.6730226874351501, + "learning_rate": 0.00012422101239430096, + "loss": 0.1508, + "step": 29440 + }, + { + "epoch": 1.1371095409089154, + "grad_norm": 2.1389102935791016, + "learning_rate": 0.00012419527137469915, + "loss": 0.3832, + "step": 29450 + }, + { + "epoch": 1.1374956562029421, + "grad_norm": 0.5423761606216431, + "learning_rate": 0.0001241695303550974, + "loss": 0.2942, + "step": 29460 + }, + { + "epoch": 1.1378817714969691, + "grad_norm": 2.6076724529266357, + "learning_rate": 0.00012414378933549557, + "loss": 0.2291, + "step": 29470 + }, + { + "epoch": 1.1382678867909959, + "grad_norm": 1.0197224617004395, + "learning_rate": 0.0001241180483158938, + "loss": 0.2309, + "step": 29480 + }, + { + "epoch": 1.1386540020850227, + "grad_norm": 1.4430413246154785, + "learning_rate": 0.00012409230729629203, + "loss": 0.258, + "step": 29490 + }, + { + "epoch": 1.1390401173790494, + "grad_norm": 1.43483304977417, + "learning_rate": 0.0001240665662766902, + "loss": 0.1931, + "step": 29500 + }, + { + "epoch": 1.1394262326730762, + "grad_norm": 0.7181301116943359, + "learning_rate": 0.00012404082525708845, + "loss": 0.2675, + "step": 29510 + }, + { + "epoch": 1.139812347967103, + "grad_norm": 2.2020421028137207, + "learning_rate": 0.00012401508423748664, + "loss": 0.5094, + "step": 29520 + }, + { + "epoch": 1.1401984632611297, + "grad_norm": 0.35750746726989746, + "learning_rate": 0.00012398934321788488, + "loss": 0.3825, + "step": 29530 + }, + { + "epoch": 1.1405845785551565, + "grad_norm": 2.1792123317718506, + "learning_rate": 0.0001239636021982831, + "loss": 0.4259, + "step": 29540 + }, + { + "epoch": 1.1409706938491833, + "grad_norm": 1.2699453830718994, + "learning_rate": 0.00012393786117868128, + "loss": 0.2524, + "step": 29550 + }, + { + "epoch": 1.1413568091432102, + "grad_norm": 4.232237339019775, + "learning_rate": 0.00012391212015907952, + "loss": 0.4191, + "step": 29560 + }, + { + "epoch": 1.141742924437237, + "grad_norm": 1.5009098052978516, + "learning_rate": 0.0001238863791394777, + "loss": 0.5748, + "step": 29570 + }, + { + "epoch": 1.1421290397312638, + "grad_norm": 0.8117336630821228, + "learning_rate": 0.00012386063811987595, + "loss": 0.2309, + "step": 29580 + }, + { + "epoch": 1.1425151550252906, + "grad_norm": 0.6417378187179565, + "learning_rate": 0.00012383489710027413, + "loss": 0.3011, + "step": 29590 + }, + { + "epoch": 1.1429012703193173, + "grad_norm": 0.19958554208278656, + "learning_rate": 0.00012380915608067237, + "loss": 0.4943, + "step": 29600 + }, + { + "epoch": 1.143287385613344, + "grad_norm": 1.7980111837387085, + "learning_rate": 0.0001237834150610706, + "loss": 0.2498, + "step": 29610 + }, + { + "epoch": 1.1436735009073709, + "grad_norm": 0.7506774663925171, + "learning_rate": 0.00012375767404146877, + "loss": 0.1622, + "step": 29620 + }, + { + "epoch": 1.1440596162013978, + "grad_norm": 0.4346953332424164, + "learning_rate": 0.000123731933021867, + "loss": 0.2608, + "step": 29630 + }, + { + "epoch": 1.1444457314954246, + "grad_norm": 1.4491907358169556, + "learning_rate": 0.0001237061920022652, + "loss": 0.4674, + "step": 29640 + }, + { + "epoch": 1.1448318467894514, + "grad_norm": 0.26328304409980774, + "learning_rate": 0.00012368045098266344, + "loss": 0.4287, + "step": 29650 + }, + { + "epoch": 1.1452179620834781, + "grad_norm": 0.6601302027702332, + "learning_rate": 0.00012365470996306163, + "loss": 0.3235, + "step": 29660 + }, + { + "epoch": 1.145604077377505, + "grad_norm": 3.9714503288269043, + "learning_rate": 0.00012362896894345987, + "loss": 0.2969, + "step": 29670 + }, + { + "epoch": 1.1459901926715317, + "grad_norm": 2.6341910362243652, + "learning_rate": 0.00012360322792385808, + "loss": 0.2771, + "step": 29680 + }, + { + "epoch": 1.1463763079655584, + "grad_norm": 0.04610513150691986, + "learning_rate": 0.00012357748690425627, + "loss": 0.2286, + "step": 29690 + }, + { + "epoch": 1.1467624232595852, + "grad_norm": 0.18892113864421844, + "learning_rate": 0.0001235517458846545, + "loss": 0.1821, + "step": 29700 + }, + { + "epoch": 1.147148538553612, + "grad_norm": 2.186973810195923, + "learning_rate": 0.0001235260048650527, + "loss": 0.2582, + "step": 29710 + }, + { + "epoch": 1.147534653847639, + "grad_norm": 0.23074299097061157, + "learning_rate": 0.00012350026384545093, + "loss": 0.3134, + "step": 29720 + }, + { + "epoch": 1.1479207691416657, + "grad_norm": 4.021410942077637, + "learning_rate": 0.00012347452282584915, + "loss": 0.3161, + "step": 29730 + }, + { + "epoch": 1.1483068844356925, + "grad_norm": 1.0262142419815063, + "learning_rate": 0.00012344878180624736, + "loss": 0.2034, + "step": 29740 + }, + { + "epoch": 1.1486929997297193, + "grad_norm": 1.4075149297714233, + "learning_rate": 0.00012342304078664557, + "loss": 0.2799, + "step": 29750 + }, + { + "epoch": 1.149079115023746, + "grad_norm": 0.9609633684158325, + "learning_rate": 0.00012339729976704376, + "loss": 0.1358, + "step": 29760 + }, + { + "epoch": 1.1494652303177728, + "grad_norm": 2.0923006534576416, + "learning_rate": 0.000123371558747442, + "loss": 0.224, + "step": 29770 + }, + { + "epoch": 1.1498513456117996, + "grad_norm": 0.4723201096057892, + "learning_rate": 0.00012334581772784019, + "loss": 0.1231, + "step": 29780 + }, + { + "epoch": 1.1502374609058266, + "grad_norm": 0.2070016860961914, + "learning_rate": 0.00012332007670823843, + "loss": 0.2998, + "step": 29790 + }, + { + "epoch": 1.1506235761998533, + "grad_norm": 0.3871285021305084, + "learning_rate": 0.00012329433568863664, + "loss": 0.2776, + "step": 29800 + }, + { + "epoch": 1.15100969149388, + "grad_norm": 0.6765030026435852, + "learning_rate": 0.00012326859466903485, + "loss": 0.6047, + "step": 29810 + }, + { + "epoch": 1.1513958067879069, + "grad_norm": 0.22907795011997223, + "learning_rate": 0.00012324285364943307, + "loss": 0.1573, + "step": 29820 + }, + { + "epoch": 1.1517819220819336, + "grad_norm": 0.3838706612586975, + "learning_rate": 0.00012321711262983128, + "loss": 0.1877, + "step": 29830 + }, + { + "epoch": 1.1521680373759604, + "grad_norm": 1.6967968940734863, + "learning_rate": 0.0001231913716102295, + "loss": 0.2314, + "step": 29840 + }, + { + "epoch": 1.1525541526699872, + "grad_norm": 0.6754477024078369, + "learning_rate": 0.00012316563059062768, + "loss": 0.1997, + "step": 29850 + }, + { + "epoch": 1.1529402679640142, + "grad_norm": 0.8980739712715149, + "learning_rate": 0.00012313988957102592, + "loss": 0.2729, + "step": 29860 + }, + { + "epoch": 1.153326383258041, + "grad_norm": 2.7968523502349854, + "learning_rate": 0.00012311414855142413, + "loss": 0.2805, + "step": 29870 + }, + { + "epoch": 1.1537124985520677, + "grad_norm": 1.5867468118667603, + "learning_rate": 0.00012308840753182235, + "loss": 0.3489, + "step": 29880 + }, + { + "epoch": 1.1540986138460945, + "grad_norm": 3.077193021774292, + "learning_rate": 0.00012306266651222056, + "loss": 0.3397, + "step": 29890 + }, + { + "epoch": 1.1544847291401212, + "grad_norm": 0.49896860122680664, + "learning_rate": 0.00012303692549261877, + "loss": 0.1215, + "step": 29900 + }, + { + "epoch": 1.154870844434148, + "grad_norm": 0.9598873257637024, + "learning_rate": 0.00012301118447301699, + "loss": 0.1618, + "step": 29910 + }, + { + "epoch": 1.1552569597281748, + "grad_norm": 1.2659916877746582, + "learning_rate": 0.0001229854434534152, + "loss": 0.1623, + "step": 29920 + }, + { + "epoch": 1.1556430750222015, + "grad_norm": 1.4009841680526733, + "learning_rate": 0.0001229597024338134, + "loss": 0.2205, + "step": 29930 + }, + { + "epoch": 1.1560291903162283, + "grad_norm": 0.5847800970077515, + "learning_rate": 0.00012293396141421163, + "loss": 0.3173, + "step": 29940 + }, + { + "epoch": 1.1564153056102553, + "grad_norm": 1.2930784225463867, + "learning_rate": 0.00012290822039460984, + "loss": 0.3052, + "step": 29950 + }, + { + "epoch": 1.156801420904282, + "grad_norm": 0.7307919859886169, + "learning_rate": 0.00012288247937500805, + "loss": 0.2293, + "step": 29960 + }, + { + "epoch": 1.1571875361983088, + "grad_norm": 0.2120385468006134, + "learning_rate": 0.00012285673835540627, + "loss": 0.2006, + "step": 29970 + }, + { + "epoch": 1.1575736514923356, + "grad_norm": 1.0932674407958984, + "learning_rate": 0.00012283099733580448, + "loss": 0.1716, + "step": 29980 + }, + { + "epoch": 1.1579597667863624, + "grad_norm": 0.7607210874557495, + "learning_rate": 0.0001228052563162027, + "loss": 0.346, + "step": 29990 + }, + { + "epoch": 1.1583458820803891, + "grad_norm": 4.455061435699463, + "learning_rate": 0.0001227795152966009, + "loss": 0.2588, + "step": 30000 + }, + { + "epoch": 1.158731997374416, + "grad_norm": 2.0812416076660156, + "learning_rate": 0.00012275377427699912, + "loss": 0.2958, + "step": 30010 + }, + { + "epoch": 1.159118112668443, + "grad_norm": 1.8883334398269653, + "learning_rate": 0.00012272803325739733, + "loss": 0.1687, + "step": 30020 + }, + { + "epoch": 1.1595042279624697, + "grad_norm": 1.3935341835021973, + "learning_rate": 0.00012270229223779555, + "loss": 0.2276, + "step": 30030 + }, + { + "epoch": 1.1598903432564964, + "grad_norm": 2.5193631649017334, + "learning_rate": 0.00012267655121819376, + "loss": 0.3598, + "step": 30040 + }, + { + "epoch": 1.1602764585505232, + "grad_norm": 1.9172290563583374, + "learning_rate": 0.00012265081019859197, + "loss": 0.3294, + "step": 30050 + }, + { + "epoch": 1.16066257384455, + "grad_norm": 1.2220836877822876, + "learning_rate": 0.00012262506917899019, + "loss": 0.2053, + "step": 30060 + }, + { + "epoch": 1.1610486891385767, + "grad_norm": 1.9804691076278687, + "learning_rate": 0.0001225993281593884, + "loss": 0.2091, + "step": 30070 + }, + { + "epoch": 1.1614348044326035, + "grad_norm": 1.2194398641586304, + "learning_rate": 0.0001225735871397866, + "loss": 0.2535, + "step": 30080 + }, + { + "epoch": 1.1618209197266305, + "grad_norm": 0.7828000783920288, + "learning_rate": 0.00012254784612018483, + "loss": 0.436, + "step": 30090 + }, + { + "epoch": 1.1622070350206573, + "grad_norm": 0.26130637526512146, + "learning_rate": 0.00012252210510058304, + "loss": 0.3351, + "step": 30100 + }, + { + "epoch": 1.162593150314684, + "grad_norm": 0.9175068736076355, + "learning_rate": 0.00012249636408098125, + "loss": 0.0682, + "step": 30110 + }, + { + "epoch": 1.1629792656087108, + "grad_norm": 0.35828933119773865, + "learning_rate": 0.00012247062306137947, + "loss": 0.3081, + "step": 30120 + }, + { + "epoch": 1.1633653809027376, + "grad_norm": 1.2477439641952515, + "learning_rate": 0.00012244488204177768, + "loss": 0.2522, + "step": 30130 + }, + { + "epoch": 1.1637514961967643, + "grad_norm": 1.0479830503463745, + "learning_rate": 0.0001224191410221759, + "loss": 0.3183, + "step": 30140 + }, + { + "epoch": 1.164137611490791, + "grad_norm": 0.9295257329940796, + "learning_rate": 0.0001223934000025741, + "loss": 0.149, + "step": 30150 + }, + { + "epoch": 1.1645237267848179, + "grad_norm": 1.4081065654754639, + "learning_rate": 0.00012236765898297232, + "loss": 0.2403, + "step": 30160 + }, + { + "epoch": 1.1649098420788446, + "grad_norm": 0.6170324087142944, + "learning_rate": 0.00012234191796337053, + "loss": 0.3526, + "step": 30170 + }, + { + "epoch": 1.1652959573728716, + "grad_norm": 3.095670461654663, + "learning_rate": 0.00012231617694376875, + "loss": 0.2873, + "step": 30180 + }, + { + "epoch": 1.1656820726668984, + "grad_norm": 1.453447937965393, + "learning_rate": 0.00012229043592416696, + "loss": 0.2753, + "step": 30190 + }, + { + "epoch": 1.1660681879609252, + "grad_norm": 1.008033275604248, + "learning_rate": 0.00012226469490456517, + "loss": 0.2926, + "step": 30200 + }, + { + "epoch": 1.166454303254952, + "grad_norm": 2.122175693511963, + "learning_rate": 0.00012223895388496339, + "loss": 0.2432, + "step": 30210 + }, + { + "epoch": 1.1668404185489787, + "grad_norm": 1.4835058450698853, + "learning_rate": 0.0001222132128653616, + "loss": 0.2891, + "step": 30220 + }, + { + "epoch": 1.1672265338430055, + "grad_norm": 1.7386225461959839, + "learning_rate": 0.00012218747184575984, + "loss": 0.2371, + "step": 30230 + }, + { + "epoch": 1.1676126491370322, + "grad_norm": 2.1769731044769287, + "learning_rate": 0.00012216173082615803, + "loss": 0.2798, + "step": 30240 + }, + { + "epoch": 1.1679987644310592, + "grad_norm": 3.3941900730133057, + "learning_rate": 0.00012213598980655624, + "loss": 0.4021, + "step": 30250 + }, + { + "epoch": 1.168384879725086, + "grad_norm": 1.099238395690918, + "learning_rate": 0.00012211024878695445, + "loss": 0.2908, + "step": 30260 + }, + { + "epoch": 1.1687709950191127, + "grad_norm": 0.6718109250068665, + "learning_rate": 0.00012208450776735267, + "loss": 0.1791, + "step": 30270 + }, + { + "epoch": 1.1691571103131395, + "grad_norm": 0.25414201617240906, + "learning_rate": 0.00012205876674775088, + "loss": 0.1322, + "step": 30280 + }, + { + "epoch": 1.1695432256071663, + "grad_norm": 2.1115262508392334, + "learning_rate": 0.00012203302572814909, + "loss": 0.3538, + "step": 30290 + }, + { + "epoch": 1.169929340901193, + "grad_norm": 2.3652501106262207, + "learning_rate": 0.00012200728470854732, + "loss": 0.3229, + "step": 30300 + }, + { + "epoch": 1.1703154561952198, + "grad_norm": 1.4749270677566528, + "learning_rate": 0.00012198154368894552, + "loss": 0.1251, + "step": 30310 + }, + { + "epoch": 1.1707015714892468, + "grad_norm": 0.6566292643547058, + "learning_rate": 0.00012195580266934375, + "loss": 0.2527, + "step": 30320 + }, + { + "epoch": 1.1710876867832736, + "grad_norm": 1.9602152109146118, + "learning_rate": 0.00012193006164974195, + "loss": 0.1851, + "step": 30330 + }, + { + "epoch": 1.1714738020773003, + "grad_norm": 1.6631299257278442, + "learning_rate": 0.00012190432063014017, + "loss": 0.4715, + "step": 30340 + }, + { + "epoch": 1.171859917371327, + "grad_norm": 1.1554430723190308, + "learning_rate": 0.00012187857961053839, + "loss": 0.361, + "step": 30350 + }, + { + "epoch": 1.1722460326653539, + "grad_norm": 2.5738513469696045, + "learning_rate": 0.00012185283859093659, + "loss": 0.3661, + "step": 30360 + }, + { + "epoch": 1.1726321479593806, + "grad_norm": 0.5713154077529907, + "learning_rate": 0.00012182709757133481, + "loss": 0.2468, + "step": 30370 + }, + { + "epoch": 1.1730182632534074, + "grad_norm": 0.7371454834938049, + "learning_rate": 0.00012180135655173301, + "loss": 0.2949, + "step": 30380 + }, + { + "epoch": 1.1734043785474342, + "grad_norm": 2.5442118644714355, + "learning_rate": 0.00012177561553213124, + "loss": 0.2305, + "step": 30390 + }, + { + "epoch": 1.173790493841461, + "grad_norm": 1.684951663017273, + "learning_rate": 0.00012174987451252944, + "loss": 0.2535, + "step": 30400 + }, + { + "epoch": 1.174176609135488, + "grad_norm": 1.6874382495880127, + "learning_rate": 0.00012172413349292767, + "loss": 0.4069, + "step": 30410 + }, + { + "epoch": 1.1745627244295147, + "grad_norm": 0.46226370334625244, + "learning_rate": 0.00012169839247332588, + "loss": 0.221, + "step": 30420 + }, + { + "epoch": 1.1749488397235415, + "grad_norm": 3.5687646865844727, + "learning_rate": 0.00012167265145372408, + "loss": 0.2749, + "step": 30430 + }, + { + "epoch": 1.1753349550175682, + "grad_norm": 2.9427647590637207, + "learning_rate": 0.0001216469104341223, + "loss": 0.2621, + "step": 30440 + }, + { + "epoch": 1.175721070311595, + "grad_norm": 1.3569320440292358, + "learning_rate": 0.0001216211694145205, + "loss": 0.1711, + "step": 30450 + }, + { + "epoch": 1.1761071856056218, + "grad_norm": 0.5906672477722168, + "learning_rate": 0.00012159542839491873, + "loss": 0.3111, + "step": 30460 + }, + { + "epoch": 1.1764933008996485, + "grad_norm": 1.7809525728225708, + "learning_rate": 0.00012156968737531693, + "loss": 0.198, + "step": 30470 + }, + { + "epoch": 1.1768794161936755, + "grad_norm": 1.5865052938461304, + "learning_rate": 0.00012154394635571516, + "loss": 0.3589, + "step": 30480 + }, + { + "epoch": 1.1772655314877023, + "grad_norm": 0.4852294623851776, + "learning_rate": 0.00012151820533611337, + "loss": 0.203, + "step": 30490 + }, + { + "epoch": 1.177651646781729, + "grad_norm": 4.437458515167236, + "learning_rate": 0.00012149246431651158, + "loss": 0.2886, + "step": 30500 + }, + { + "epoch": 1.1780377620757558, + "grad_norm": 2.204751491546631, + "learning_rate": 0.0001214667232969098, + "loss": 0.313, + "step": 30510 + }, + { + "epoch": 1.1784238773697826, + "grad_norm": 0.9356504678726196, + "learning_rate": 0.000121440982277308, + "loss": 0.2813, + "step": 30520 + }, + { + "epoch": 1.1788099926638094, + "grad_norm": 0.06744952499866486, + "learning_rate": 0.00012141524125770622, + "loss": 0.2009, + "step": 30530 + }, + { + "epoch": 1.1791961079578361, + "grad_norm": 0.49779242277145386, + "learning_rate": 0.00012138950023810445, + "loss": 0.1822, + "step": 30540 + }, + { + "epoch": 1.1795822232518631, + "grad_norm": 1.1115593910217285, + "learning_rate": 0.00012136375921850265, + "loss": 0.4164, + "step": 30550 + }, + { + "epoch": 1.17996833854589, + "grad_norm": 0.2939944267272949, + "learning_rate": 0.00012133801819890086, + "loss": 0.1356, + "step": 30560 + }, + { + "epoch": 1.1803544538399167, + "grad_norm": 0.15905381739139557, + "learning_rate": 0.00012131227717929908, + "loss": 0.3479, + "step": 30570 + }, + { + "epoch": 1.1807405691339434, + "grad_norm": 3.146277666091919, + "learning_rate": 0.00012128653615969729, + "loss": 0.2722, + "step": 30580 + }, + { + "epoch": 1.1811266844279702, + "grad_norm": 0.60884690284729, + "learning_rate": 0.00012126079514009549, + "loss": 0.1778, + "step": 30590 + }, + { + "epoch": 1.181512799721997, + "grad_norm": 0.053204573690891266, + "learning_rate": 0.00012123505412049372, + "loss": 0.1397, + "step": 30600 + }, + { + "epoch": 1.1818989150160237, + "grad_norm": 1.2908227443695068, + "learning_rate": 0.00012120931310089194, + "loss": 0.1552, + "step": 30610 + }, + { + "epoch": 1.1822850303100505, + "grad_norm": 1.489575743675232, + "learning_rate": 0.00012118357208129014, + "loss": 0.2965, + "step": 30620 + }, + { + "epoch": 1.1826711456040773, + "grad_norm": 2.4103262424468994, + "learning_rate": 0.00012115783106168837, + "loss": 0.2947, + "step": 30630 + }, + { + "epoch": 1.1830572608981043, + "grad_norm": 0.3685878813266754, + "learning_rate": 0.00012113209004208657, + "loss": 0.2144, + "step": 30640 + }, + { + "epoch": 1.183443376192131, + "grad_norm": 0.29953858256340027, + "learning_rate": 0.00012110634902248478, + "loss": 0.2281, + "step": 30650 + }, + { + "epoch": 1.1838294914861578, + "grad_norm": 0.4792311191558838, + "learning_rate": 0.00012108060800288298, + "loss": 0.2727, + "step": 30660 + }, + { + "epoch": 1.1842156067801846, + "grad_norm": 0.17143972218036652, + "learning_rate": 0.00012105486698328121, + "loss": 0.2096, + "step": 30670 + }, + { + "epoch": 1.1846017220742113, + "grad_norm": 0.43678683042526245, + "learning_rate": 0.00012102912596367944, + "loss": 0.3818, + "step": 30680 + }, + { + "epoch": 1.184987837368238, + "grad_norm": 2.0908610820770264, + "learning_rate": 0.00012100338494407764, + "loss": 0.2496, + "step": 30690 + }, + { + "epoch": 1.1853739526622649, + "grad_norm": 1.5331153869628906, + "learning_rate": 0.00012097764392447586, + "loss": 0.2772, + "step": 30700 + }, + { + "epoch": 1.1857600679562919, + "grad_norm": 0.6948639154434204, + "learning_rate": 0.00012095190290487406, + "loss": 0.3896, + "step": 30710 + }, + { + "epoch": 1.1861461832503186, + "grad_norm": 2.3802030086517334, + "learning_rate": 0.00012092616188527228, + "loss": 0.1998, + "step": 30720 + }, + { + "epoch": 1.1865322985443454, + "grad_norm": 2.1358511447906494, + "learning_rate": 0.0001209004208656705, + "loss": 0.2865, + "step": 30730 + }, + { + "epoch": 1.1869184138383722, + "grad_norm": 0.8761110305786133, + "learning_rate": 0.0001208746798460687, + "loss": 0.2574, + "step": 30740 + }, + { + "epoch": 1.187304529132399, + "grad_norm": 0.6293401718139648, + "learning_rate": 0.00012084893882646693, + "loss": 0.4006, + "step": 30750 + }, + { + "epoch": 1.1876906444264257, + "grad_norm": 2.4534378051757812, + "learning_rate": 0.00012082319780686513, + "loss": 0.2281, + "step": 30760 + }, + { + "epoch": 1.1880767597204525, + "grad_norm": 0.38392508029937744, + "learning_rate": 0.00012079745678726336, + "loss": 0.3055, + "step": 30770 + }, + { + "epoch": 1.1884628750144794, + "grad_norm": 1.278108835220337, + "learning_rate": 0.00012077171576766156, + "loss": 0.2376, + "step": 30780 + }, + { + "epoch": 1.1888489903085062, + "grad_norm": 0.2575186789035797, + "learning_rate": 0.00012074597474805977, + "loss": 0.256, + "step": 30790 + }, + { + "epoch": 1.189235105602533, + "grad_norm": 4.043684959411621, + "learning_rate": 0.000120720233728458, + "loss": 0.3601, + "step": 30800 + }, + { + "epoch": 1.1896212208965598, + "grad_norm": 0.2561960220336914, + "learning_rate": 0.0001206944927088562, + "loss": 0.2698, + "step": 30810 + }, + { + "epoch": 1.1900073361905865, + "grad_norm": 0.3477579355239868, + "learning_rate": 0.00012066875168925442, + "loss": 0.3442, + "step": 30820 + }, + { + "epoch": 1.1903934514846133, + "grad_norm": 1.0357879400253296, + "learning_rate": 0.00012064301066965262, + "loss": 0.2561, + "step": 30830 + }, + { + "epoch": 1.19077956677864, + "grad_norm": 0.14653460681438446, + "learning_rate": 0.00012061726965005085, + "loss": 0.3072, + "step": 30840 + }, + { + "epoch": 1.1911656820726668, + "grad_norm": 1.2516767978668213, + "learning_rate": 0.00012059152863044906, + "loss": 0.4203, + "step": 30850 + }, + { + "epoch": 1.1915517973666936, + "grad_norm": 0.41542065143585205, + "learning_rate": 0.00012056578761084726, + "loss": 0.2018, + "step": 30860 + }, + { + "epoch": 1.1919379126607206, + "grad_norm": 1.7187318801879883, + "learning_rate": 0.00012054004659124549, + "loss": 0.097, + "step": 30870 + }, + { + "epoch": 1.1923240279547473, + "grad_norm": 0.3913117051124573, + "learning_rate": 0.00012051430557164369, + "loss": 0.1655, + "step": 30880 + }, + { + "epoch": 1.1927101432487741, + "grad_norm": 3.708519697189331, + "learning_rate": 0.00012048856455204192, + "loss": 0.2315, + "step": 30890 + }, + { + "epoch": 1.1930962585428009, + "grad_norm": 1.506726861000061, + "learning_rate": 0.00012046282353244012, + "loss": 0.1712, + "step": 30900 + }, + { + "epoch": 1.1934823738368276, + "grad_norm": 1.5069276094436646, + "learning_rate": 0.00012043708251283834, + "loss": 0.399, + "step": 30910 + }, + { + "epoch": 1.1938684891308544, + "grad_norm": 0.09606973081827164, + "learning_rate": 0.00012041134149323656, + "loss": 0.1226, + "step": 30920 + }, + { + "epoch": 1.1942546044248812, + "grad_norm": 1.7927659749984741, + "learning_rate": 0.00012038560047363476, + "loss": 0.0975, + "step": 30930 + }, + { + "epoch": 1.1946407197189082, + "grad_norm": 2.2988364696502686, + "learning_rate": 0.00012035985945403298, + "loss": 0.2742, + "step": 30940 + }, + { + "epoch": 1.195026835012935, + "grad_norm": 0.4412599205970764, + "learning_rate": 0.00012033411843443118, + "loss": 0.2008, + "step": 30950 + }, + { + "epoch": 1.1954129503069617, + "grad_norm": 3.227698564529419, + "learning_rate": 0.00012030837741482941, + "loss": 0.2797, + "step": 30960 + }, + { + "epoch": 1.1957990656009885, + "grad_norm": 1.3904643058776855, + "learning_rate": 0.00012028263639522761, + "loss": 0.1636, + "step": 30970 + }, + { + "epoch": 1.1961851808950152, + "grad_norm": 1.6495708227157593, + "learning_rate": 0.00012025689537562584, + "loss": 0.2768, + "step": 30980 + }, + { + "epoch": 1.196571296189042, + "grad_norm": 0.3142000436782837, + "learning_rate": 0.00012023115435602405, + "loss": 0.1797, + "step": 30990 + }, + { + "epoch": 1.1969574114830688, + "grad_norm": 1.562090516090393, + "learning_rate": 0.00012020541333642225, + "loss": 0.3175, + "step": 31000 + }, + { + "epoch": 1.1973435267770955, + "grad_norm": 1.3837881088256836, + "learning_rate": 0.00012017967231682048, + "loss": 0.3137, + "step": 31010 + }, + { + "epoch": 1.1977296420711223, + "grad_norm": 1.1495468616485596, + "learning_rate": 0.00012015393129721868, + "loss": 0.232, + "step": 31020 + }, + { + "epoch": 1.1981157573651493, + "grad_norm": 1.562514305114746, + "learning_rate": 0.0001201281902776169, + "loss": 0.3455, + "step": 31030 + }, + { + "epoch": 1.198501872659176, + "grad_norm": 1.626610517501831, + "learning_rate": 0.00012010244925801513, + "loss": 0.2382, + "step": 31040 + }, + { + "epoch": 1.1988879879532028, + "grad_norm": 2.5863585472106934, + "learning_rate": 0.00012007670823841333, + "loss": 0.288, + "step": 31050 + }, + { + "epoch": 1.1992741032472296, + "grad_norm": 0.9751160740852356, + "learning_rate": 0.00012005096721881154, + "loss": 0.3236, + "step": 31060 + }, + { + "epoch": 1.1996602185412564, + "grad_norm": 1.7428686618804932, + "learning_rate": 0.00012002522619920974, + "loss": 0.3707, + "step": 31070 + }, + { + "epoch": 1.2000463338352831, + "grad_norm": 1.0963342189788818, + "learning_rate": 0.00011999948517960797, + "loss": 0.2991, + "step": 31080 + }, + { + "epoch": 1.20043244912931, + "grad_norm": 2.0942695140838623, + "learning_rate": 0.00011997374416000617, + "loss": 0.2611, + "step": 31090 + }, + { + "epoch": 1.200818564423337, + "grad_norm": 0.11989244073629379, + "learning_rate": 0.0001199480031404044, + "loss": 0.2071, + "step": 31100 + }, + { + "epoch": 1.2012046797173637, + "grad_norm": 0.44892773032188416, + "learning_rate": 0.00011992226212080262, + "loss": 0.2105, + "step": 31110 + }, + { + "epoch": 1.2015907950113904, + "grad_norm": 0.0861232578754425, + "learning_rate": 0.00011989652110120082, + "loss": 0.1906, + "step": 31120 + }, + { + "epoch": 1.2019769103054172, + "grad_norm": 1.9162683486938477, + "learning_rate": 0.00011987078008159904, + "loss": 0.1929, + "step": 31130 + }, + { + "epoch": 1.202363025599444, + "grad_norm": 1.3750224113464355, + "learning_rate": 0.00011984503906199725, + "loss": 0.3072, + "step": 31140 + }, + { + "epoch": 1.2027491408934707, + "grad_norm": 0.49282577633857727, + "learning_rate": 0.00011981929804239546, + "loss": 0.3082, + "step": 31150 + }, + { + "epoch": 1.2031352561874975, + "grad_norm": 1.0578473806381226, + "learning_rate": 0.00011979355702279366, + "loss": 0.2498, + "step": 31160 + }, + { + "epoch": 1.2035213714815245, + "grad_norm": 0.6070756316184998, + "learning_rate": 0.00011976781600319189, + "loss": 0.2324, + "step": 31170 + }, + { + "epoch": 1.2039074867755513, + "grad_norm": 3.2918875217437744, + "learning_rate": 0.00011974207498359012, + "loss": 0.4218, + "step": 31180 + }, + { + "epoch": 1.204293602069578, + "grad_norm": 0.3284684419631958, + "learning_rate": 0.00011971633396398832, + "loss": 0.2713, + "step": 31190 + }, + { + "epoch": 1.2046797173636048, + "grad_norm": 0.8733110427856445, + "learning_rate": 0.00011969059294438653, + "loss": 0.328, + "step": 31200 + }, + { + "epoch": 1.2050658326576316, + "grad_norm": 1.787016749382019, + "learning_rate": 0.00011966485192478474, + "loss": 0.2115, + "step": 31210 + }, + { + "epoch": 1.2054519479516583, + "grad_norm": 0.7539357542991638, + "learning_rate": 0.00011963911090518296, + "loss": 0.3585, + "step": 31220 + }, + { + "epoch": 1.205838063245685, + "grad_norm": 0.32271450757980347, + "learning_rate": 0.00011961336988558118, + "loss": 0.2472, + "step": 31230 + }, + { + "epoch": 1.2062241785397119, + "grad_norm": 0.85898756980896, + "learning_rate": 0.00011958762886597938, + "loss": 0.2367, + "step": 31240 + }, + { + "epoch": 1.2066102938337386, + "grad_norm": 0.6451173424720764, + "learning_rate": 0.00011956188784637761, + "loss": 0.2309, + "step": 31250 + }, + { + "epoch": 1.2069964091277656, + "grad_norm": 2.1043314933776855, + "learning_rate": 0.00011953614682677581, + "loss": 0.316, + "step": 31260 + }, + { + "epoch": 1.2073825244217924, + "grad_norm": 2.9022254943847656, + "learning_rate": 0.00011951040580717404, + "loss": 0.3161, + "step": 31270 + }, + { + "epoch": 1.2077686397158192, + "grad_norm": 0.7839256525039673, + "learning_rate": 0.00011948466478757224, + "loss": 0.4014, + "step": 31280 + }, + { + "epoch": 1.208154755009846, + "grad_norm": 1.024190902709961, + "learning_rate": 0.00011945892376797045, + "loss": 0.2333, + "step": 31290 + }, + { + "epoch": 1.2085408703038727, + "grad_norm": 1.433605670928955, + "learning_rate": 0.00011943318274836868, + "loss": 0.2014, + "step": 31300 + }, + { + "epoch": 1.2089269855978995, + "grad_norm": 0.7208861112594604, + "learning_rate": 0.00011940744172876688, + "loss": 0.1786, + "step": 31310 + }, + { + "epoch": 1.2093131008919262, + "grad_norm": 2.5464839935302734, + "learning_rate": 0.0001193817007091651, + "loss": 0.3646, + "step": 31320 + }, + { + "epoch": 1.2096992161859532, + "grad_norm": 1.6425135135650635, + "learning_rate": 0.0001193559596895633, + "loss": 0.2603, + "step": 31330 + }, + { + "epoch": 1.21008533147998, + "grad_norm": 1.7573461532592773, + "learning_rate": 0.00011933021866996153, + "loss": 0.315, + "step": 31340 + }, + { + "epoch": 1.2104714467740068, + "grad_norm": 1.884445071220398, + "learning_rate": 0.00011930447765035974, + "loss": 0.2326, + "step": 31350 + }, + { + "epoch": 1.2108575620680335, + "grad_norm": 1.2781214714050293, + "learning_rate": 0.00011927873663075794, + "loss": 0.2248, + "step": 31360 + }, + { + "epoch": 1.2112436773620603, + "grad_norm": 0.5210689902305603, + "learning_rate": 0.00011925299561115617, + "loss": 0.2517, + "step": 31370 + }, + { + "epoch": 1.211629792656087, + "grad_norm": 0.4265996515750885, + "learning_rate": 0.00011922725459155437, + "loss": 0.257, + "step": 31380 + }, + { + "epoch": 1.2120159079501138, + "grad_norm": 0.7833511233329773, + "learning_rate": 0.0001192015135719526, + "loss": 0.3117, + "step": 31390 + }, + { + "epoch": 1.2124020232441408, + "grad_norm": 5.922067165374756, + "learning_rate": 0.0001191757725523508, + "loss": 0.2949, + "step": 31400 + }, + { + "epoch": 1.2127881385381676, + "grad_norm": 3.2208797931671143, + "learning_rate": 0.00011915003153274902, + "loss": 0.3616, + "step": 31410 + }, + { + "epoch": 1.2131742538321943, + "grad_norm": 1.6283681392669678, + "learning_rate": 0.00011912429051314724, + "loss": 0.1951, + "step": 31420 + }, + { + "epoch": 1.2135603691262211, + "grad_norm": 1.2450298070907593, + "learning_rate": 0.00011909854949354544, + "loss": 0.2313, + "step": 31430 + }, + { + "epoch": 1.2139464844202479, + "grad_norm": 0.8394368290901184, + "learning_rate": 0.00011907280847394366, + "loss": 0.2993, + "step": 31440 + }, + { + "epoch": 1.2143325997142747, + "grad_norm": 0.6099762320518494, + "learning_rate": 0.00011904706745434186, + "loss": 0.1527, + "step": 31450 + }, + { + "epoch": 1.2147187150083014, + "grad_norm": 0.7743226885795593, + "learning_rate": 0.00011902132643474009, + "loss": 0.2608, + "step": 31460 + }, + { + "epoch": 1.2151048303023282, + "grad_norm": 3.343604326248169, + "learning_rate": 0.00011899558541513829, + "loss": 0.3031, + "step": 31470 + }, + { + "epoch": 1.215490945596355, + "grad_norm": 1.016830563545227, + "learning_rate": 0.00011896984439553652, + "loss": 0.2146, + "step": 31480 + }, + { + "epoch": 1.215877060890382, + "grad_norm": 1.0264688730239868, + "learning_rate": 0.00011894410337593473, + "loss": 0.2905, + "step": 31490 + }, + { + "epoch": 1.2162631761844087, + "grad_norm": 1.5718705654144287, + "learning_rate": 0.00011891836235633293, + "loss": 0.2928, + "step": 31500 + }, + { + "epoch": 1.2166492914784355, + "grad_norm": 0.9886181950569153, + "learning_rate": 0.00011889262133673116, + "loss": 0.2129, + "step": 31510 + }, + { + "epoch": 1.2170354067724622, + "grad_norm": 0.6496497392654419, + "learning_rate": 0.00011886688031712936, + "loss": 0.2388, + "step": 31520 + }, + { + "epoch": 1.217421522066489, + "grad_norm": 2.0419809818267822, + "learning_rate": 0.00011884113929752758, + "loss": 0.1835, + "step": 31530 + }, + { + "epoch": 1.2178076373605158, + "grad_norm": 0.41370299458503723, + "learning_rate": 0.00011881539827792581, + "loss": 0.3099, + "step": 31540 + }, + { + "epoch": 1.2181937526545425, + "grad_norm": 0.6608594655990601, + "learning_rate": 0.00011878965725832401, + "loss": 0.4235, + "step": 31550 + }, + { + "epoch": 1.2185798679485695, + "grad_norm": 0.4206163287162781, + "learning_rate": 0.00011876391623872222, + "loss": 0.2384, + "step": 31560 + }, + { + "epoch": 1.2189659832425963, + "grad_norm": 0.2976207137107849, + "learning_rate": 0.00011873817521912042, + "loss": 0.4792, + "step": 31570 + }, + { + "epoch": 1.219352098536623, + "grad_norm": 2.237607002258301, + "learning_rate": 0.00011871243419951865, + "loss": 0.4694, + "step": 31580 + }, + { + "epoch": 1.2197382138306498, + "grad_norm": 2.1085827350616455, + "learning_rate": 0.00011868669317991685, + "loss": 0.3475, + "step": 31590 + }, + { + "epoch": 1.2201243291246766, + "grad_norm": 1.5770317316055298, + "learning_rate": 0.00011866095216031508, + "loss": 0.3676, + "step": 31600 + }, + { + "epoch": 1.2205104444187034, + "grad_norm": 1.7337925434112549, + "learning_rate": 0.0001186352111407133, + "loss": 0.2663, + "step": 31610 + }, + { + "epoch": 1.2208965597127301, + "grad_norm": 0.23846319317817688, + "learning_rate": 0.0001186094701211115, + "loss": 0.1916, + "step": 31620 + }, + { + "epoch": 1.2212826750067571, + "grad_norm": 1.1316229104995728, + "learning_rate": 0.00011858372910150972, + "loss": 0.1924, + "step": 31630 + }, + { + "epoch": 1.221668790300784, + "grad_norm": 0.7642451524734497, + "learning_rate": 0.00011855798808190792, + "loss": 0.3786, + "step": 31640 + }, + { + "epoch": 1.2220549055948107, + "grad_norm": 1.975182056427002, + "learning_rate": 0.00011853224706230614, + "loss": 0.2646, + "step": 31650 + }, + { + "epoch": 1.2224410208888374, + "grad_norm": 1.5172406435012817, + "learning_rate": 0.00011850650604270437, + "loss": 0.2824, + "step": 31660 + }, + { + "epoch": 1.2228271361828642, + "grad_norm": 1.5113869905471802, + "learning_rate": 0.00011848076502310257, + "loss": 0.2959, + "step": 31670 + }, + { + "epoch": 1.223213251476891, + "grad_norm": 2.380364418029785, + "learning_rate": 0.0001184550240035008, + "loss": 0.2425, + "step": 31680 + }, + { + "epoch": 1.2235993667709177, + "grad_norm": 0.11588902771472931, + "learning_rate": 0.000118429282983899, + "loss": 0.1665, + "step": 31690 + }, + { + "epoch": 1.2239854820649445, + "grad_norm": 2.045466184616089, + "learning_rate": 0.00011840354196429721, + "loss": 0.2478, + "step": 31700 + }, + { + "epoch": 1.2243715973589713, + "grad_norm": 4.201779365539551, + "learning_rate": 0.00011837780094469542, + "loss": 0.3295, + "step": 31710 + }, + { + "epoch": 1.2247577126529983, + "grad_norm": 1.1080553531646729, + "learning_rate": 0.00011835205992509364, + "loss": 0.2102, + "step": 31720 + }, + { + "epoch": 1.225143827947025, + "grad_norm": 0.3263179659843445, + "learning_rate": 0.00011832631890549186, + "loss": 0.2822, + "step": 31730 + }, + { + "epoch": 1.2255299432410518, + "grad_norm": 0.7899855971336365, + "learning_rate": 0.00011830057788589006, + "loss": 0.1462, + "step": 31740 + }, + { + "epoch": 1.2259160585350786, + "grad_norm": 1.330854058265686, + "learning_rate": 0.00011827483686628829, + "loss": 0.3883, + "step": 31750 + }, + { + "epoch": 1.2263021738291053, + "grad_norm": 0.40446966886520386, + "learning_rate": 0.00011824909584668649, + "loss": 0.1592, + "step": 31760 + }, + { + "epoch": 1.226688289123132, + "grad_norm": 0.49187901616096497, + "learning_rate": 0.0001182233548270847, + "loss": 0.0698, + "step": 31770 + }, + { + "epoch": 1.2270744044171589, + "grad_norm": 4.227303504943848, + "learning_rate": 0.00011819761380748292, + "loss": 0.3945, + "step": 31780 + }, + { + "epoch": 1.2274605197111859, + "grad_norm": 2.520578622817993, + "learning_rate": 0.00011817187278788113, + "loss": 0.4552, + "step": 31790 + }, + { + "epoch": 1.2278466350052126, + "grad_norm": 0.9486772418022156, + "learning_rate": 0.00011814613176827936, + "loss": 0.232, + "step": 31800 + }, + { + "epoch": 1.2282327502992394, + "grad_norm": 0.8336694836616516, + "learning_rate": 0.00011812039074867756, + "loss": 0.183, + "step": 31810 + }, + { + "epoch": 1.2286188655932662, + "grad_norm": 1.576710820198059, + "learning_rate": 0.00011809464972907578, + "loss": 0.4116, + "step": 31820 + }, + { + "epoch": 1.229004980887293, + "grad_norm": 0.2388358861207962, + "learning_rate": 0.00011806890870947398, + "loss": 0.3509, + "step": 31830 + }, + { + "epoch": 1.2293910961813197, + "grad_norm": 1.054880142211914, + "learning_rate": 0.00011804316768987221, + "loss": 0.1669, + "step": 31840 + }, + { + "epoch": 1.2297772114753465, + "grad_norm": 0.7959414720535278, + "learning_rate": 0.00011801742667027042, + "loss": 0.2073, + "step": 31850 + }, + { + "epoch": 1.2301633267693735, + "grad_norm": 2.480940103530884, + "learning_rate": 0.00011799168565066862, + "loss": 0.3589, + "step": 31860 + }, + { + "epoch": 1.2305494420634002, + "grad_norm": 0.9381177425384521, + "learning_rate": 0.00011796594463106685, + "loss": 0.2576, + "step": 31870 + }, + { + "epoch": 1.230935557357427, + "grad_norm": 9.548282623291016, + "learning_rate": 0.00011794020361146505, + "loss": 0.1755, + "step": 31880 + }, + { + "epoch": 1.2313216726514538, + "grad_norm": 0.4922927916049957, + "learning_rate": 0.00011791446259186328, + "loss": 0.2173, + "step": 31890 + }, + { + "epoch": 1.2317077879454805, + "grad_norm": 1.2846627235412598, + "learning_rate": 0.00011788872157226148, + "loss": 0.2039, + "step": 31900 + }, + { + "epoch": 1.2320939032395073, + "grad_norm": 0.9321876764297485, + "learning_rate": 0.0001178629805526597, + "loss": 0.2418, + "step": 31910 + }, + { + "epoch": 1.232480018533534, + "grad_norm": 1.373787760734558, + "learning_rate": 0.00011783723953305792, + "loss": 0.3065, + "step": 31920 + }, + { + "epoch": 1.2328661338275608, + "grad_norm": 1.915208339691162, + "learning_rate": 0.00011781149851345612, + "loss": 0.292, + "step": 31930 + }, + { + "epoch": 1.2332522491215876, + "grad_norm": 1.533010482788086, + "learning_rate": 0.00011778575749385434, + "loss": 0.2234, + "step": 31940 + }, + { + "epoch": 1.2336383644156146, + "grad_norm": 0.5740505456924438, + "learning_rate": 0.00011776001647425254, + "loss": 0.2513, + "step": 31950 + }, + { + "epoch": 1.2340244797096414, + "grad_norm": 3.164320230484009, + "learning_rate": 0.00011773427545465077, + "loss": 0.4233, + "step": 31960 + }, + { + "epoch": 1.2344105950036681, + "grad_norm": 0.8309218287467957, + "learning_rate": 0.00011770853443504897, + "loss": 0.4629, + "step": 31970 + }, + { + "epoch": 1.2347967102976949, + "grad_norm": 2.0942423343658447, + "learning_rate": 0.0001176827934154472, + "loss": 0.3381, + "step": 31980 + }, + { + "epoch": 1.2351828255917217, + "grad_norm": 0.4961155354976654, + "learning_rate": 0.00011765705239584541, + "loss": 0.1265, + "step": 31990 + }, + { + "epoch": 1.2355689408857484, + "grad_norm": 1.6777870655059814, + "learning_rate": 0.00011763131137624361, + "loss": 0.2503, + "step": 32000 + }, + { + "epoch": 1.2359550561797752, + "grad_norm": 0.6016990542411804, + "learning_rate": 0.00011760557035664184, + "loss": 0.3944, + "step": 32010 + }, + { + "epoch": 1.2363411714738022, + "grad_norm": 1.9123533964157104, + "learning_rate": 0.00011757982933704004, + "loss": 0.463, + "step": 32020 + }, + { + "epoch": 1.236727286767829, + "grad_norm": 1.5328630208969116, + "learning_rate": 0.00011755408831743826, + "loss": 0.3368, + "step": 32030 + }, + { + "epoch": 1.2371134020618557, + "grad_norm": 1.0984220504760742, + "learning_rate": 0.00011752834729783649, + "loss": 0.1987, + "step": 32040 + }, + { + "epoch": 1.2374995173558825, + "grad_norm": 3.1570327281951904, + "learning_rate": 0.00011750260627823469, + "loss": 0.454, + "step": 32050 + }, + { + "epoch": 1.2378856326499093, + "grad_norm": 3.7589874267578125, + "learning_rate": 0.0001174768652586329, + "loss": 0.2953, + "step": 32060 + }, + { + "epoch": 1.238271747943936, + "grad_norm": 1.3289847373962402, + "learning_rate": 0.0001174511242390311, + "loss": 0.3058, + "step": 32070 + }, + { + "epoch": 1.2386578632379628, + "grad_norm": 0.9872431755065918, + "learning_rate": 0.00011742538321942933, + "loss": 0.1312, + "step": 32080 + }, + { + "epoch": 1.2390439785319898, + "grad_norm": 1.799133062362671, + "learning_rate": 0.00011739964219982753, + "loss": 0.3271, + "step": 32090 + }, + { + "epoch": 1.2394300938260165, + "grad_norm": 0.13501858711242676, + "learning_rate": 0.00011737390118022576, + "loss": 0.4462, + "step": 32100 + }, + { + "epoch": 1.2398162091200433, + "grad_norm": 1.53009033203125, + "learning_rate": 0.00011734816016062398, + "loss": 0.1311, + "step": 32110 + }, + { + "epoch": 1.24020232441407, + "grad_norm": 0.5737781524658203, + "learning_rate": 0.00011732241914102218, + "loss": 0.1705, + "step": 32120 + }, + { + "epoch": 1.2405884397080968, + "grad_norm": 2.095947027206421, + "learning_rate": 0.0001172966781214204, + "loss": 0.2195, + "step": 32130 + }, + { + "epoch": 1.2409745550021236, + "grad_norm": 2.2158639430999756, + "learning_rate": 0.0001172709371018186, + "loss": 0.2964, + "step": 32140 + }, + { + "epoch": 1.2413606702961504, + "grad_norm": 0.3154670000076294, + "learning_rate": 0.00011724519608221682, + "loss": 0.245, + "step": 32150 + }, + { + "epoch": 1.2417467855901771, + "grad_norm": 1.4467942714691162, + "learning_rate": 0.00011721945506261505, + "loss": 0.3469, + "step": 32160 + }, + { + "epoch": 1.242132900884204, + "grad_norm": 0.24252896010875702, + "learning_rate": 0.00011719371404301325, + "loss": 0.2217, + "step": 32170 + }, + { + "epoch": 1.242519016178231, + "grad_norm": 2.4256465435028076, + "learning_rate": 0.00011716797302341148, + "loss": 0.4707, + "step": 32180 + }, + { + "epoch": 1.2429051314722577, + "grad_norm": 0.5917278528213501, + "learning_rate": 0.00011714223200380968, + "loss": 0.2284, + "step": 32190 + }, + { + "epoch": 1.2432912467662844, + "grad_norm": 1.2977256774902344, + "learning_rate": 0.00011711649098420789, + "loss": 0.3612, + "step": 32200 + }, + { + "epoch": 1.2436773620603112, + "grad_norm": 1.296558141708374, + "learning_rate": 0.00011709074996460609, + "loss": 0.2231, + "step": 32210 + }, + { + "epoch": 1.244063477354338, + "grad_norm": 1.6559596061706543, + "learning_rate": 0.00011706500894500432, + "loss": 0.2636, + "step": 32220 + }, + { + "epoch": 1.2444495926483647, + "grad_norm": 1.6342560052871704, + "learning_rate": 0.00011703926792540254, + "loss": 0.1276, + "step": 32230 + }, + { + "epoch": 1.2448357079423915, + "grad_norm": 1.1173146963119507, + "learning_rate": 0.00011701352690580074, + "loss": 0.1719, + "step": 32240 + }, + { + "epoch": 1.2452218232364185, + "grad_norm": 0.29269275069236755, + "learning_rate": 0.00011698778588619897, + "loss": 0.3323, + "step": 32250 + }, + { + "epoch": 1.2456079385304453, + "grad_norm": 2.518568515777588, + "learning_rate": 0.00011696204486659717, + "loss": 0.3302, + "step": 32260 + }, + { + "epoch": 1.245994053824472, + "grad_norm": 2.535940647125244, + "learning_rate": 0.00011693630384699538, + "loss": 0.1488, + "step": 32270 + }, + { + "epoch": 1.2463801691184988, + "grad_norm": 0.384199321269989, + "learning_rate": 0.00011691056282739358, + "loss": 0.0957, + "step": 32280 + }, + { + "epoch": 1.2467662844125256, + "grad_norm": 3.7460570335388184, + "learning_rate": 0.00011688482180779181, + "loss": 0.311, + "step": 32290 + }, + { + "epoch": 1.2471523997065523, + "grad_norm": 1.285152792930603, + "learning_rate": 0.00011685908078819004, + "loss": 0.1872, + "step": 32300 + }, + { + "epoch": 1.247538515000579, + "grad_norm": 0.6118050217628479, + "learning_rate": 0.00011683333976858823, + "loss": 0.1635, + "step": 32310 + }, + { + "epoch": 1.2479246302946059, + "grad_norm": 2.9890856742858887, + "learning_rate": 0.00011680759874898646, + "loss": 0.2773, + "step": 32320 + }, + { + "epoch": 1.2483107455886326, + "grad_norm": 0.15415504574775696, + "learning_rate": 0.00011678185772938466, + "loss": 0.3014, + "step": 32330 + }, + { + "epoch": 1.2486968608826596, + "grad_norm": 1.2717432975769043, + "learning_rate": 0.00011675611670978287, + "loss": 0.2442, + "step": 32340 + }, + { + "epoch": 1.2490829761766864, + "grad_norm": 0.19885335862636566, + "learning_rate": 0.0001167303756901811, + "loss": 0.1894, + "step": 32350 + }, + { + "epoch": 1.2494690914707132, + "grad_norm": 1.425176978111267, + "learning_rate": 0.0001167046346705793, + "loss": 0.3008, + "step": 32360 + }, + { + "epoch": 1.24985520676474, + "grad_norm": 1.9033544063568115, + "learning_rate": 0.00011667889365097753, + "loss": 0.3072, + "step": 32370 + }, + { + "epoch": 1.2502413220587667, + "grad_norm": 0.6677396297454834, + "learning_rate": 0.00011665315263137573, + "loss": 0.2378, + "step": 32380 + }, + { + "epoch": 1.2506274373527935, + "grad_norm": 0.4491410255432129, + "learning_rate": 0.00011662741161177396, + "loss": 0.1071, + "step": 32390 + }, + { + "epoch": 1.2510135526468202, + "grad_norm": 2.7697031497955322, + "learning_rate": 0.00011660167059217215, + "loss": 0.3276, + "step": 32400 + }, + { + "epoch": 1.2513996679408472, + "grad_norm": 1.4374775886535645, + "learning_rate": 0.00011657592957257037, + "loss": 0.2459, + "step": 32410 + }, + { + "epoch": 1.251785783234874, + "grad_norm": 1.5245740413665771, + "learning_rate": 0.0001165501885529686, + "loss": 0.3841, + "step": 32420 + }, + { + "epoch": 1.2521718985289008, + "grad_norm": 0.5069687366485596, + "learning_rate": 0.0001165244475333668, + "loss": 0.1425, + "step": 32430 + }, + { + "epoch": 1.2525580138229275, + "grad_norm": 0.6181765198707581, + "learning_rate": 0.00011649870651376502, + "loss": 0.4184, + "step": 32440 + }, + { + "epoch": 1.2529441291169543, + "grad_norm": 2.007375717163086, + "learning_rate": 0.00011647296549416322, + "loss": 0.1754, + "step": 32450 + }, + { + "epoch": 1.253330244410981, + "grad_norm": 0.27261993288993835, + "learning_rate": 0.00011644722447456145, + "loss": 0.3608, + "step": 32460 + }, + { + "epoch": 1.2537163597050078, + "grad_norm": 0.3452116549015045, + "learning_rate": 0.00011642148345495965, + "loss": 0.1913, + "step": 32470 + }, + { + "epoch": 1.2541024749990348, + "grad_norm": 2.553738594055176, + "learning_rate": 0.00011639574243535787, + "loss": 0.2221, + "step": 32480 + }, + { + "epoch": 1.2544885902930616, + "grad_norm": 0.795706033706665, + "learning_rate": 0.00011637000141575609, + "loss": 0.2348, + "step": 32490 + }, + { + "epoch": 1.2548747055870884, + "grad_norm": 0.519059956073761, + "learning_rate": 0.00011634426039615429, + "loss": 0.2874, + "step": 32500 + }, + { + "epoch": 1.2552608208811151, + "grad_norm": 2.7050392627716064, + "learning_rate": 0.00011631851937655251, + "loss": 0.3821, + "step": 32510 + }, + { + "epoch": 1.255646936175142, + "grad_norm": 2.09036922454834, + "learning_rate": 0.00011629277835695071, + "loss": 0.2562, + "step": 32520 + }, + { + "epoch": 1.2560330514691687, + "grad_norm": 1.156534194946289, + "learning_rate": 0.00011626703733734894, + "loss": 0.1373, + "step": 32530 + }, + { + "epoch": 1.2564191667631954, + "grad_norm": 2.5214719772338867, + "learning_rate": 0.00011624129631774715, + "loss": 0.2955, + "step": 32540 + }, + { + "epoch": 1.2568052820572224, + "grad_norm": 2.71368408203125, + "learning_rate": 0.00011621555529814537, + "loss": 0.2832, + "step": 32550 + }, + { + "epoch": 1.257191397351249, + "grad_norm": 1.488390564918518, + "learning_rate": 0.00011618981427854358, + "loss": 0.2369, + "step": 32560 + }, + { + "epoch": 1.257577512645276, + "grad_norm": 1.7705353498458862, + "learning_rate": 0.00011616407325894178, + "loss": 0.4351, + "step": 32570 + }, + { + "epoch": 1.2579636279393027, + "grad_norm": 4.3224406242370605, + "learning_rate": 0.00011613833223934001, + "loss": 0.3365, + "step": 32580 + }, + { + "epoch": 1.2583497432333295, + "grad_norm": 0.9157351851463318, + "learning_rate": 0.00011611259121973821, + "loss": 0.2882, + "step": 32590 + }, + { + "epoch": 1.2587358585273563, + "grad_norm": 2.9030823707580566, + "learning_rate": 0.00011608685020013643, + "loss": 0.343, + "step": 32600 + }, + { + "epoch": 1.259121973821383, + "grad_norm": 0.807424783706665, + "learning_rate": 0.00011606110918053466, + "loss": 0.2413, + "step": 32610 + }, + { + "epoch": 1.2595080891154098, + "grad_norm": 1.8564451932907104, + "learning_rate": 0.00011603536816093286, + "loss": 0.5403, + "step": 32620 + }, + { + "epoch": 1.2598942044094366, + "grad_norm": 0.8169263005256653, + "learning_rate": 0.00011600962714133107, + "loss": 0.3522, + "step": 32630 + }, + { + "epoch": 1.2602803197034635, + "grad_norm": 0.9638017416000366, + "learning_rate": 0.00011598388612172927, + "loss": 0.3295, + "step": 32640 + }, + { + "epoch": 1.2606664349974903, + "grad_norm": 1.8613898754119873, + "learning_rate": 0.0001159581451021275, + "loss": 0.3107, + "step": 32650 + }, + { + "epoch": 1.261052550291517, + "grad_norm": 2.582638740539551, + "learning_rate": 0.00011593240408252573, + "loss": 0.2767, + "step": 32660 + }, + { + "epoch": 1.2614386655855439, + "grad_norm": 1.8227603435516357, + "learning_rate": 0.00011590666306292393, + "loss": 0.2324, + "step": 32670 + }, + { + "epoch": 1.2618247808795706, + "grad_norm": 0.3827721178531647, + "learning_rate": 0.00011588092204332215, + "loss": 0.4926, + "step": 32680 + }, + { + "epoch": 1.2622108961735974, + "grad_norm": 1.5523405075073242, + "learning_rate": 0.00011585518102372035, + "loss": 0.4475, + "step": 32690 + }, + { + "epoch": 1.2625970114676242, + "grad_norm": 0.25254619121551514, + "learning_rate": 0.00011582944000411857, + "loss": 0.2333, + "step": 32700 + }, + { + "epoch": 1.2629831267616511, + "grad_norm": 0.5530809164047241, + "learning_rate": 0.00011580369898451677, + "loss": 0.2781, + "step": 32710 + }, + { + "epoch": 1.2633692420556777, + "grad_norm": 0.3202857971191406, + "learning_rate": 0.000115777957964915, + "loss": 0.2596, + "step": 32720 + }, + { + "epoch": 1.2637553573497047, + "grad_norm": 0.25663653016090393, + "learning_rate": 0.00011575221694531322, + "loss": 0.2096, + "step": 32730 + }, + { + "epoch": 1.2641414726437314, + "grad_norm": 2.423585891723633, + "learning_rate": 0.00011572647592571142, + "loss": 0.3064, + "step": 32740 + }, + { + "epoch": 1.2645275879377582, + "grad_norm": 0.2529740631580353, + "learning_rate": 0.00011570073490610965, + "loss": 0.1426, + "step": 32750 + }, + { + "epoch": 1.264913703231785, + "grad_norm": 0.5238136053085327, + "learning_rate": 0.00011567499388650785, + "loss": 0.1598, + "step": 32760 + }, + { + "epoch": 1.2652998185258117, + "grad_norm": 0.4495049715042114, + "learning_rate": 0.00011564925286690606, + "loss": 0.1712, + "step": 32770 + }, + { + "epoch": 1.2656859338198387, + "grad_norm": 0.3863711953163147, + "learning_rate": 0.00011562351184730426, + "loss": 0.1015, + "step": 32780 + }, + { + "epoch": 1.2660720491138653, + "grad_norm": 1.3139948844909668, + "learning_rate": 0.00011559777082770249, + "loss": 0.3982, + "step": 32790 + }, + { + "epoch": 1.2664581644078923, + "grad_norm": 0.40500277280807495, + "learning_rate": 0.00011557202980810071, + "loss": 0.4211, + "step": 32800 + }, + { + "epoch": 1.266844279701919, + "grad_norm": 0.9749487042427063, + "learning_rate": 0.00011554628878849891, + "loss": 0.2773, + "step": 32810 + }, + { + "epoch": 1.2672303949959458, + "grad_norm": 1.1194144487380981, + "learning_rate": 0.00011552054776889714, + "loss": 0.2771, + "step": 32820 + }, + { + "epoch": 1.2676165102899726, + "grad_norm": 1.4098531007766724, + "learning_rate": 0.00011549480674929534, + "loss": 0.2044, + "step": 32830 + }, + { + "epoch": 1.2680026255839993, + "grad_norm": 1.3184692859649658, + "learning_rate": 0.00011546906572969355, + "loss": 0.2633, + "step": 32840 + }, + { + "epoch": 1.268388740878026, + "grad_norm": 0.9776495099067688, + "learning_rate": 0.00011544332471009178, + "loss": 0.3212, + "step": 32850 + }, + { + "epoch": 1.2687748561720529, + "grad_norm": 1.1577869653701782, + "learning_rate": 0.00011541758369048998, + "loss": 0.4271, + "step": 32860 + }, + { + "epoch": 1.2691609714660799, + "grad_norm": 2.2742063999176025, + "learning_rate": 0.00011539184267088821, + "loss": 0.2044, + "step": 32870 + }, + { + "epoch": 1.2695470867601066, + "grad_norm": 1.2156758308410645, + "learning_rate": 0.00011536610165128641, + "loss": 0.2394, + "step": 32880 + }, + { + "epoch": 1.2699332020541334, + "grad_norm": 1.9160290956497192, + "learning_rate": 0.00011534036063168463, + "loss": 0.3733, + "step": 32890 + }, + { + "epoch": 1.2703193173481602, + "grad_norm": 1.3484338521957397, + "learning_rate": 0.00011531461961208283, + "loss": 0.3946, + "step": 32900 + }, + { + "epoch": 1.270705432642187, + "grad_norm": 1.0480244159698486, + "learning_rate": 0.00011528887859248105, + "loss": 0.166, + "step": 32910 + }, + { + "epoch": 1.2710915479362137, + "grad_norm": 1.1980483531951904, + "learning_rate": 0.00011526313757287927, + "loss": 0.3134, + "step": 32920 + }, + { + "epoch": 1.2714776632302405, + "grad_norm": 1.2777379751205444, + "learning_rate": 0.00011523739655327747, + "loss": 0.2615, + "step": 32930 + }, + { + "epoch": 1.2718637785242675, + "grad_norm": 1.559792399406433, + "learning_rate": 0.0001152116555336757, + "loss": 0.2996, + "step": 32940 + }, + { + "epoch": 1.272249893818294, + "grad_norm": 1.9589471817016602, + "learning_rate": 0.0001151859145140739, + "loss": 0.2683, + "step": 32950 + }, + { + "epoch": 1.272636009112321, + "grad_norm": 0.23336051404476166, + "learning_rate": 0.00011516017349447213, + "loss": 0.3306, + "step": 32960 + }, + { + "epoch": 1.2730221244063478, + "grad_norm": 1.5898360013961792, + "learning_rate": 0.00011513443247487034, + "loss": 0.2891, + "step": 32970 + }, + { + "epoch": 1.2734082397003745, + "grad_norm": 0.5046207904815674, + "learning_rate": 0.00011510869145526854, + "loss": 0.1843, + "step": 32980 + }, + { + "epoch": 1.2737943549944013, + "grad_norm": 0.16588473320007324, + "learning_rate": 0.00011508295043566677, + "loss": 0.1065, + "step": 32990 + }, + { + "epoch": 1.274180470288428, + "grad_norm": 1.5499800443649292, + "learning_rate": 0.00011505720941606497, + "loss": 0.2795, + "step": 33000 + }, + { + "epoch": 1.274566585582455, + "grad_norm": 0.5019121766090393, + "learning_rate": 0.0001150314683964632, + "loss": 0.1857, + "step": 33010 + }, + { + "epoch": 1.2749527008764816, + "grad_norm": 0.258952260017395, + "learning_rate": 0.0001150057273768614, + "loss": 0.11, + "step": 33020 + }, + { + "epoch": 1.2753388161705086, + "grad_norm": 0.8540348410606384, + "learning_rate": 0.00011497998635725962, + "loss": 0.1852, + "step": 33030 + }, + { + "epoch": 1.2757249314645354, + "grad_norm": 0.08689398318529129, + "learning_rate": 0.00011495424533765783, + "loss": 0.2225, + "step": 33040 + }, + { + "epoch": 1.2761110467585621, + "grad_norm": 0.42253822088241577, + "learning_rate": 0.00011492850431805605, + "loss": 0.0751, + "step": 33050 + }, + { + "epoch": 1.276497162052589, + "grad_norm": 1.2964017391204834, + "learning_rate": 0.00011490276329845426, + "loss": 0.2384, + "step": 33060 + }, + { + "epoch": 1.2768832773466157, + "grad_norm": 0.5337836146354675, + "learning_rate": 0.00011487702227885246, + "loss": 0.1415, + "step": 33070 + }, + { + "epoch": 1.2772693926406424, + "grad_norm": 2.7771682739257812, + "learning_rate": 0.00011485128125925069, + "loss": 0.378, + "step": 33080 + }, + { + "epoch": 1.2776555079346692, + "grad_norm": 1.5107232332229614, + "learning_rate": 0.00011482554023964889, + "loss": 0.2482, + "step": 33090 + }, + { + "epoch": 1.2780416232286962, + "grad_norm": 0.6885499358177185, + "learning_rate": 0.00011479979922004711, + "loss": 0.2122, + "step": 33100 + }, + { + "epoch": 1.278427738522723, + "grad_norm": 0.9016557335853577, + "learning_rate": 0.00011477405820044533, + "loss": 0.2841, + "step": 33110 + }, + { + "epoch": 1.2788138538167497, + "grad_norm": 1.9532525539398193, + "learning_rate": 0.00011474831718084354, + "loss": 0.2281, + "step": 33120 + }, + { + "epoch": 1.2791999691107765, + "grad_norm": 2.1078782081604004, + "learning_rate": 0.00011472257616124175, + "loss": 0.2836, + "step": 33130 + }, + { + "epoch": 1.2795860844048033, + "grad_norm": 0.19830390810966492, + "learning_rate": 0.00011469683514163995, + "loss": 0.388, + "step": 33140 + }, + { + "epoch": 1.27997219969883, + "grad_norm": 0.17538850009441376, + "learning_rate": 0.00011467109412203818, + "loss": 0.3274, + "step": 33150 + }, + { + "epoch": 1.2803583149928568, + "grad_norm": 0.7402139902114868, + "learning_rate": 0.00011464535310243641, + "loss": 0.1979, + "step": 33160 + }, + { + "epoch": 1.2807444302868838, + "grad_norm": 0.2097146362066269, + "learning_rate": 0.00011461961208283461, + "loss": 0.2464, + "step": 33170 + }, + { + "epoch": 1.2811305455809103, + "grad_norm": 1.2441083192825317, + "learning_rate": 0.00011459387106323283, + "loss": 0.266, + "step": 33180 + }, + { + "epoch": 1.2815166608749373, + "grad_norm": 2.518852710723877, + "learning_rate": 0.00011456813004363103, + "loss": 0.253, + "step": 33190 + }, + { + "epoch": 1.281902776168964, + "grad_norm": 0.8078998327255249, + "learning_rate": 0.00011454238902402925, + "loss": 0.2361, + "step": 33200 + }, + { + "epoch": 1.2822888914629909, + "grad_norm": 1.2297371625900269, + "learning_rate": 0.00011451664800442745, + "loss": 0.1974, + "step": 33210 + }, + { + "epoch": 1.2826750067570176, + "grad_norm": 0.4303855895996094, + "learning_rate": 0.00011449090698482567, + "loss": 0.3563, + "step": 33220 + }, + { + "epoch": 1.2830611220510444, + "grad_norm": 1.3215210437774658, + "learning_rate": 0.0001144651659652239, + "loss": 0.2818, + "step": 33230 + }, + { + "epoch": 1.2834472373450712, + "grad_norm": 1.546265959739685, + "learning_rate": 0.0001144394249456221, + "loss": 0.5778, + "step": 33240 + }, + { + "epoch": 1.283833352639098, + "grad_norm": 0.8895953297615051, + "learning_rate": 0.00011441368392602033, + "loss": 0.2485, + "step": 33250 + }, + { + "epoch": 1.284219467933125, + "grad_norm": 0.7534870505332947, + "learning_rate": 0.00011438794290641853, + "loss": 0.2825, + "step": 33260 + }, + { + "epoch": 1.2846055832271517, + "grad_norm": 0.052820973098278046, + "learning_rate": 0.00011436220188681674, + "loss": 0.2191, + "step": 33270 + }, + { + "epoch": 1.2849916985211784, + "grad_norm": 0.9264475107192993, + "learning_rate": 0.00011433646086721494, + "loss": 0.181, + "step": 33280 + }, + { + "epoch": 1.2853778138152052, + "grad_norm": 0.2128441333770752, + "learning_rate": 0.00011431071984761317, + "loss": 0.1819, + "step": 33290 + }, + { + "epoch": 1.285763929109232, + "grad_norm": 0.5400950312614441, + "learning_rate": 0.0001142849788280114, + "loss": 0.4316, + "step": 33300 + }, + { + "epoch": 1.2861500444032588, + "grad_norm": 0.8033271431922913, + "learning_rate": 0.00011425923780840959, + "loss": 0.2146, + "step": 33310 + }, + { + "epoch": 1.2865361596972855, + "grad_norm": 2.012575149536133, + "learning_rate": 0.00011423349678880782, + "loss": 0.4335, + "step": 33320 + }, + { + "epoch": 1.2869222749913125, + "grad_norm": 0.7352376580238342, + "learning_rate": 0.00011420775576920602, + "loss": 0.2124, + "step": 33330 + }, + { + "epoch": 1.2873083902853393, + "grad_norm": 0.769036591053009, + "learning_rate": 0.00011418201474960423, + "loss": 0.3602, + "step": 33340 + }, + { + "epoch": 1.287694505579366, + "grad_norm": 0.250592976808548, + "learning_rate": 0.00011415627373000246, + "loss": 0.1692, + "step": 33350 + }, + { + "epoch": 1.2880806208733928, + "grad_norm": 2.43820858001709, + "learning_rate": 0.00011413053271040066, + "loss": 0.2777, + "step": 33360 + }, + { + "epoch": 1.2884667361674196, + "grad_norm": 1.3179954290390015, + "learning_rate": 0.00011410479169079889, + "loss": 0.1794, + "step": 33370 + }, + { + "epoch": 1.2888528514614463, + "grad_norm": 1.0040466785430908, + "learning_rate": 0.00011407905067119709, + "loss": 0.3037, + "step": 33380 + }, + { + "epoch": 1.2892389667554731, + "grad_norm": 5.296288013458252, + "learning_rate": 0.00011405330965159531, + "loss": 0.2904, + "step": 33390 + }, + { + "epoch": 1.2896250820495, + "grad_norm": 0.4267273247241974, + "learning_rate": 0.00011402756863199351, + "loss": 0.2263, + "step": 33400 + }, + { + "epoch": 1.2900111973435266, + "grad_norm": 0.8817713260650635, + "learning_rate": 0.00011400182761239173, + "loss": 0.2715, + "step": 33410 + }, + { + "epoch": 1.2903973126375536, + "grad_norm": 2.7891275882720947, + "learning_rate": 0.00011397608659278995, + "loss": 0.2781, + "step": 33420 + }, + { + "epoch": 1.2907834279315804, + "grad_norm": 0.3013952374458313, + "learning_rate": 0.00011395034557318815, + "loss": 0.2563, + "step": 33430 + }, + { + "epoch": 1.2911695432256072, + "grad_norm": 1.766413927078247, + "learning_rate": 0.00011392460455358638, + "loss": 0.1854, + "step": 33440 + }, + { + "epoch": 1.291555658519634, + "grad_norm": 0.25331103801727295, + "learning_rate": 0.00011389886353398458, + "loss": 0.1206, + "step": 33450 + }, + { + "epoch": 1.2919417738136607, + "grad_norm": 0.036400288343429565, + "learning_rate": 0.0001138731225143828, + "loss": 0.4707, + "step": 33460 + }, + { + "epoch": 1.2923278891076875, + "grad_norm": 1.5620888471603394, + "learning_rate": 0.00011384738149478102, + "loss": 0.3312, + "step": 33470 + }, + { + "epoch": 1.2927140044017142, + "grad_norm": 0.6670392155647278, + "learning_rate": 0.00011382164047517922, + "loss": 0.2341, + "step": 33480 + }, + { + "epoch": 1.2931001196957412, + "grad_norm": 2.3108737468719482, + "learning_rate": 0.00011379589945557745, + "loss": 0.3843, + "step": 33490 + }, + { + "epoch": 1.293486234989768, + "grad_norm": 0.8025147318840027, + "learning_rate": 0.00011377015843597565, + "loss": 0.1982, + "step": 33500 + }, + { + "epoch": 1.2938723502837948, + "grad_norm": 1.7835719585418701, + "learning_rate": 0.00011374441741637387, + "loss": 0.3285, + "step": 33510 + }, + { + "epoch": 1.2942584655778215, + "grad_norm": 2.041508913040161, + "learning_rate": 0.00011371867639677207, + "loss": 0.2044, + "step": 33520 + }, + { + "epoch": 1.2946445808718483, + "grad_norm": 1.103378415107727, + "learning_rate": 0.0001136929353771703, + "loss": 0.1682, + "step": 33530 + }, + { + "epoch": 1.295030696165875, + "grad_norm": 0.057376351207494736, + "learning_rate": 0.00011366719435756851, + "loss": 0.1642, + "step": 33540 + }, + { + "epoch": 1.2954168114599018, + "grad_norm": 0.6539410948753357, + "learning_rate": 0.00011364145333796671, + "loss": 0.1549, + "step": 33550 + }, + { + "epoch": 1.2958029267539288, + "grad_norm": 1.250543236732483, + "learning_rate": 0.00011361571231836494, + "loss": 0.3764, + "step": 33560 + }, + { + "epoch": 1.2961890420479556, + "grad_norm": 0.23697887361049652, + "learning_rate": 0.00011358997129876314, + "loss": 0.3999, + "step": 33570 + }, + { + "epoch": 1.2965751573419824, + "grad_norm": 0.9318505525588989, + "learning_rate": 0.00011356423027916137, + "loss": 0.4156, + "step": 33580 + }, + { + "epoch": 1.2969612726360091, + "grad_norm": 1.3910777568817139, + "learning_rate": 0.00011353848925955957, + "loss": 0.3455, + "step": 33590 + }, + { + "epoch": 1.297347387930036, + "grad_norm": 1.6764451265335083, + "learning_rate": 0.00011351274823995779, + "loss": 0.1884, + "step": 33600 + }, + { + "epoch": 1.2977335032240627, + "grad_norm": 0.9300051927566528, + "learning_rate": 0.000113487007220356, + "loss": 0.119, + "step": 33610 + }, + { + "epoch": 1.2981196185180894, + "grad_norm": 2.447462558746338, + "learning_rate": 0.00011346126620075422, + "loss": 0.4403, + "step": 33620 + }, + { + "epoch": 1.2985057338121164, + "grad_norm": 1.216407060623169, + "learning_rate": 0.00011343552518115243, + "loss": 0.2415, + "step": 33630 + }, + { + "epoch": 1.298891849106143, + "grad_norm": 2.968648910522461, + "learning_rate": 0.00011340978416155063, + "loss": 0.2899, + "step": 33640 + }, + { + "epoch": 1.29927796440017, + "grad_norm": 0.6649970412254333, + "learning_rate": 0.00011338404314194886, + "loss": 0.3809, + "step": 33650 + }, + { + "epoch": 1.2996640796941967, + "grad_norm": 1.7277917861938477, + "learning_rate": 0.00011335830212234709, + "loss": 0.3308, + "step": 33660 + }, + { + "epoch": 1.3000501949882235, + "grad_norm": 1.3269709348678589, + "learning_rate": 0.00011333256110274529, + "loss": 0.3682, + "step": 33670 + }, + { + "epoch": 1.3004363102822503, + "grad_norm": 0.20609407126903534, + "learning_rate": 0.0001133068200831435, + "loss": 0.1379, + "step": 33680 + }, + { + "epoch": 1.300822425576277, + "grad_norm": 0.6592215299606323, + "learning_rate": 0.00011328107906354171, + "loss": 0.2746, + "step": 33690 + }, + { + "epoch": 1.3012085408703038, + "grad_norm": 1.903635859489441, + "learning_rate": 0.00011325533804393993, + "loss": 0.4729, + "step": 33700 + }, + { + "epoch": 1.3015946561643306, + "grad_norm": 0.8432504534721375, + "learning_rate": 0.00011322959702433813, + "loss": 0.2835, + "step": 33710 + }, + { + "epoch": 1.3019807714583576, + "grad_norm": 0.9862542152404785, + "learning_rate": 0.00011320385600473635, + "loss": 0.1823, + "step": 33720 + }, + { + "epoch": 1.3023668867523843, + "grad_norm": 3.845738649368286, + "learning_rate": 0.00011317811498513458, + "loss": 0.2542, + "step": 33730 + }, + { + "epoch": 1.302753002046411, + "grad_norm": 0.6317747235298157, + "learning_rate": 0.00011315237396553278, + "loss": 0.22, + "step": 33740 + }, + { + "epoch": 1.3031391173404379, + "grad_norm": 2.5221354961395264, + "learning_rate": 0.000113126632945931, + "loss": 0.2253, + "step": 33750 + }, + { + "epoch": 1.3035252326344646, + "grad_norm": 1.3326247930526733, + "learning_rate": 0.0001131008919263292, + "loss": 0.2021, + "step": 33760 + }, + { + "epoch": 1.3039113479284914, + "grad_norm": 1.148047685623169, + "learning_rate": 0.00011307515090672742, + "loss": 0.3987, + "step": 33770 + }, + { + "epoch": 1.3042974632225182, + "grad_norm": 0.19721268117427826, + "learning_rate": 0.00011304940988712562, + "loss": 0.2642, + "step": 33780 + }, + { + "epoch": 1.3046835785165452, + "grad_norm": 1.4060617685317993, + "learning_rate": 0.00011302366886752385, + "loss": 0.2736, + "step": 33790 + }, + { + "epoch": 1.305069693810572, + "grad_norm": 1.0736548900604248, + "learning_rate": 0.00011299792784792207, + "loss": 0.2448, + "step": 33800 + }, + { + "epoch": 1.3054558091045987, + "grad_norm": 4.352476119995117, + "learning_rate": 0.00011297218682832027, + "loss": 0.383, + "step": 33810 + }, + { + "epoch": 1.3058419243986255, + "grad_norm": 0.2249228060245514, + "learning_rate": 0.0001129464458087185, + "loss": 0.14, + "step": 33820 + }, + { + "epoch": 1.3062280396926522, + "grad_norm": 0.4820781946182251, + "learning_rate": 0.0001129207047891167, + "loss": 0.248, + "step": 33830 + }, + { + "epoch": 1.306614154986679, + "grad_norm": 2.2983391284942627, + "learning_rate": 0.00011289496376951491, + "loss": 0.2608, + "step": 33840 + }, + { + "epoch": 1.3070002702807058, + "grad_norm": 1.3315671682357788, + "learning_rate": 0.00011286922274991314, + "loss": 0.1727, + "step": 33850 + }, + { + "epoch": 1.3073863855747327, + "grad_norm": 2.060299873352051, + "learning_rate": 0.00011284348173031134, + "loss": 0.3097, + "step": 33860 + }, + { + "epoch": 1.3077725008687593, + "grad_norm": 2.096285581588745, + "learning_rate": 0.00011281774071070957, + "loss": 0.2305, + "step": 33870 + }, + { + "epoch": 1.3081586161627863, + "grad_norm": 0.4997636675834656, + "learning_rate": 0.00011279199969110777, + "loss": 0.1993, + "step": 33880 + }, + { + "epoch": 1.308544731456813, + "grad_norm": 0.58636474609375, + "learning_rate": 0.00011276625867150599, + "loss": 0.1873, + "step": 33890 + }, + { + "epoch": 1.3089308467508398, + "grad_norm": 0.9128592610359192, + "learning_rate": 0.00011274051765190419, + "loss": 0.1885, + "step": 33900 + }, + { + "epoch": 1.3093169620448666, + "grad_norm": 2.228043794631958, + "learning_rate": 0.0001127147766323024, + "loss": 0.3649, + "step": 33910 + }, + { + "epoch": 1.3097030773388934, + "grad_norm": 1.069002389907837, + "learning_rate": 0.00011268903561270063, + "loss": 0.5454, + "step": 33920 + }, + { + "epoch": 1.3100891926329201, + "grad_norm": 0.6207597851753235, + "learning_rate": 0.00011266329459309883, + "loss": 0.2329, + "step": 33930 + }, + { + "epoch": 1.3104753079269469, + "grad_norm": 1.262247920036316, + "learning_rate": 0.00011263755357349706, + "loss": 0.3437, + "step": 33940 + }, + { + "epoch": 1.3108614232209739, + "grad_norm": 1.7429994344711304, + "learning_rate": 0.00011261181255389526, + "loss": 0.228, + "step": 33950 + }, + { + "epoch": 1.3112475385150006, + "grad_norm": 0.646900475025177, + "learning_rate": 0.00011258607153429349, + "loss": 0.3739, + "step": 33960 + }, + { + "epoch": 1.3116336538090274, + "grad_norm": 1.8228782415390015, + "learning_rate": 0.0001125603305146917, + "loss": 0.2325, + "step": 33970 + }, + { + "epoch": 1.3120197691030542, + "grad_norm": 3.539228916168213, + "learning_rate": 0.0001125345894950899, + "loss": 0.194, + "step": 33980 + }, + { + "epoch": 1.312405884397081, + "grad_norm": 1.2801135778427124, + "learning_rate": 0.00011250884847548813, + "loss": 0.3069, + "step": 33990 + }, + { + "epoch": 1.3127919996911077, + "grad_norm": 3.6265695095062256, + "learning_rate": 0.00011248310745588633, + "loss": 0.3113, + "step": 34000 + }, + { + "epoch": 1.3131781149851345, + "grad_norm": 0.07370063662528992, + "learning_rate": 0.00011245736643628455, + "loss": 0.1449, + "step": 34010 + }, + { + "epoch": 1.3135642302791615, + "grad_norm": 1.0295637845993042, + "learning_rate": 0.00011243162541668275, + "loss": 0.22, + "step": 34020 + }, + { + "epoch": 1.313950345573188, + "grad_norm": 0.8803662061691284, + "learning_rate": 0.00011240588439708098, + "loss": 0.1368, + "step": 34030 + }, + { + "epoch": 1.314336460867215, + "grad_norm": 1.6597707271575928, + "learning_rate": 0.00011238014337747919, + "loss": 0.3038, + "step": 34040 + }, + { + "epoch": 1.3147225761612418, + "grad_norm": 2.115492343902588, + "learning_rate": 0.00011235440235787739, + "loss": 0.1754, + "step": 34050 + }, + { + "epoch": 1.3151086914552685, + "grad_norm": 0.8143919706344604, + "learning_rate": 0.00011232866133827562, + "loss": 0.3764, + "step": 34060 + }, + { + "epoch": 1.3154948067492953, + "grad_norm": 0.14369767904281616, + "learning_rate": 0.00011230292031867382, + "loss": 0.1742, + "step": 34070 + }, + { + "epoch": 1.315880922043322, + "grad_norm": 1.0129845142364502, + "learning_rate": 0.00011227717929907205, + "loss": 0.1458, + "step": 34080 + }, + { + "epoch": 1.316267037337349, + "grad_norm": 2.7300291061401367, + "learning_rate": 0.00011225143827947025, + "loss": 0.3939, + "step": 34090 + }, + { + "epoch": 1.3166531526313756, + "grad_norm": 0.20205609500408173, + "learning_rate": 0.00011222569725986847, + "loss": 0.29, + "step": 34100 + }, + { + "epoch": 1.3170392679254026, + "grad_norm": 1.8928464651107788, + "learning_rate": 0.00011219995624026669, + "loss": 0.1742, + "step": 34110 + }, + { + "epoch": 1.3174253832194294, + "grad_norm": 0.2639687955379486, + "learning_rate": 0.00011217421522066488, + "loss": 0.1745, + "step": 34120 + }, + { + "epoch": 1.3178114985134561, + "grad_norm": 0.5906389355659485, + "learning_rate": 0.00011214847420106311, + "loss": 0.2134, + "step": 34130 + }, + { + "epoch": 1.318197613807483, + "grad_norm": 0.9190629720687866, + "learning_rate": 0.00011212273318146131, + "loss": 0.2547, + "step": 34140 + }, + { + "epoch": 1.3185837291015097, + "grad_norm": 0.5749151110649109, + "learning_rate": 0.00011209699216185954, + "loss": 0.1688, + "step": 34150 + }, + { + "epoch": 1.3189698443955364, + "grad_norm": 0.82295823097229, + "learning_rate": 0.00011207125114225777, + "loss": 0.2884, + "step": 34160 + }, + { + "epoch": 1.3193559596895632, + "grad_norm": 0.07816460728645325, + "learning_rate": 0.00011204551012265597, + "loss": 0.2418, + "step": 34170 + }, + { + "epoch": 1.3197420749835902, + "grad_norm": 0.6417407393455505, + "learning_rate": 0.00011201976910305418, + "loss": 0.2557, + "step": 34180 + }, + { + "epoch": 1.320128190277617, + "grad_norm": 6.093267440795898, + "learning_rate": 0.00011199402808345238, + "loss": 0.3088, + "step": 34190 + }, + { + "epoch": 1.3205143055716437, + "grad_norm": 1.8861887454986572, + "learning_rate": 0.0001119682870638506, + "loss": 0.2204, + "step": 34200 + }, + { + "epoch": 1.3209004208656705, + "grad_norm": 2.3272714614868164, + "learning_rate": 0.0001119425460442488, + "loss": 0.2236, + "step": 34210 + }, + { + "epoch": 1.3212865361596973, + "grad_norm": 0.9608810544013977, + "learning_rate": 0.00011191680502464703, + "loss": 0.1897, + "step": 34220 + }, + { + "epoch": 1.321672651453724, + "grad_norm": 1.2157350778579712, + "learning_rate": 0.00011189106400504526, + "loss": 0.1526, + "step": 34230 + }, + { + "epoch": 1.3220587667477508, + "grad_norm": 1.6684671640396118, + "learning_rate": 0.00011186532298544346, + "loss": 0.3394, + "step": 34240 + }, + { + "epoch": 1.3224448820417778, + "grad_norm": 2.0432374477386475, + "learning_rate": 0.00011183958196584167, + "loss": 0.2183, + "step": 34250 + }, + { + "epoch": 1.3228309973358043, + "grad_norm": 0.9436892867088318, + "learning_rate": 0.00011181384094623988, + "loss": 0.2947, + "step": 34260 + }, + { + "epoch": 1.3232171126298313, + "grad_norm": 0.23260092735290527, + "learning_rate": 0.0001117880999266381, + "loss": 0.114, + "step": 34270 + }, + { + "epoch": 1.323603227923858, + "grad_norm": 1.2291594743728638, + "learning_rate": 0.0001117623589070363, + "loss": 0.3145, + "step": 34280 + }, + { + "epoch": 1.3239893432178849, + "grad_norm": 0.41411107778549194, + "learning_rate": 0.00011173661788743452, + "loss": 0.2937, + "step": 34290 + }, + { + "epoch": 1.3243754585119116, + "grad_norm": 2.354405164718628, + "learning_rate": 0.00011171087686783275, + "loss": 0.3933, + "step": 34300 + }, + { + "epoch": 1.3247615738059384, + "grad_norm": 2.6997978687286377, + "learning_rate": 0.00011168513584823095, + "loss": 0.1494, + "step": 34310 + }, + { + "epoch": 1.3251476890999654, + "grad_norm": 2.8430919647216797, + "learning_rate": 0.00011165939482862916, + "loss": 0.2869, + "step": 34320 + }, + { + "epoch": 1.325533804393992, + "grad_norm": 1.1737356185913086, + "learning_rate": 0.00011163365380902738, + "loss": 0.2792, + "step": 34330 + }, + { + "epoch": 1.325919919688019, + "grad_norm": 4.123973846435547, + "learning_rate": 0.00011160791278942559, + "loss": 0.5211, + "step": 34340 + }, + { + "epoch": 1.3263060349820457, + "grad_norm": 0.8862038850784302, + "learning_rate": 0.00011158217176982382, + "loss": 0.2976, + "step": 34350 + }, + { + "epoch": 1.3266921502760725, + "grad_norm": 1.8690590858459473, + "learning_rate": 0.00011155643075022202, + "loss": 0.2485, + "step": 34360 + }, + { + "epoch": 1.3270782655700992, + "grad_norm": 2.885589599609375, + "learning_rate": 0.00011153068973062024, + "loss": 0.3084, + "step": 34370 + }, + { + "epoch": 1.327464380864126, + "grad_norm": 0.9898788928985596, + "learning_rate": 0.00011150494871101844, + "loss": 0.261, + "step": 34380 + }, + { + "epoch": 1.3278504961581528, + "grad_norm": 0.6879653930664062, + "learning_rate": 0.00011147920769141667, + "loss": 0.2082, + "step": 34390 + }, + { + "epoch": 1.3282366114521795, + "grad_norm": 1.2619003057479858, + "learning_rate": 0.00011145346667181487, + "loss": 0.2402, + "step": 34400 + }, + { + "epoch": 1.3286227267462065, + "grad_norm": 1.1212007999420166, + "learning_rate": 0.00011142772565221308, + "loss": 0.3525, + "step": 34410 + }, + { + "epoch": 1.3290088420402333, + "grad_norm": 1.8431956768035889, + "learning_rate": 0.00011140198463261131, + "loss": 0.2212, + "step": 34420 + }, + { + "epoch": 1.32939495733426, + "grad_norm": 0.6185423731803894, + "learning_rate": 0.00011137624361300951, + "loss": 0.2455, + "step": 34430 + }, + { + "epoch": 1.3297810726282868, + "grad_norm": 2.3791301250457764, + "learning_rate": 0.00011135050259340774, + "loss": 0.1763, + "step": 34440 + }, + { + "epoch": 1.3301671879223136, + "grad_norm": 0.4928603768348694, + "learning_rate": 0.00011132476157380594, + "loss": 0.2381, + "step": 34450 + }, + { + "epoch": 1.3305533032163404, + "grad_norm": 1.5636029243469238, + "learning_rate": 0.00011129902055420416, + "loss": 0.1368, + "step": 34460 + }, + { + "epoch": 1.3309394185103671, + "grad_norm": 0.9425283074378967, + "learning_rate": 0.00011127327953460238, + "loss": 0.2825, + "step": 34470 + }, + { + "epoch": 1.3313255338043941, + "grad_norm": 1.2257115840911865, + "learning_rate": 0.00011124753851500058, + "loss": 0.2547, + "step": 34480 + }, + { + "epoch": 1.3317116490984207, + "grad_norm": 0.9416170716285706, + "learning_rate": 0.0001112217974953988, + "loss": 0.2766, + "step": 34490 + }, + { + "epoch": 1.3320977643924476, + "grad_norm": 0.5123847126960754, + "learning_rate": 0.000111196056475797, + "loss": 0.4733, + "step": 34500 + }, + { + "epoch": 1.3324838796864744, + "grad_norm": 1.5581384897232056, + "learning_rate": 0.00011117031545619523, + "loss": 0.1597, + "step": 34510 + }, + { + "epoch": 1.3328699949805012, + "grad_norm": 2.377333879470825, + "learning_rate": 0.00011114457443659343, + "loss": 0.209, + "step": 34520 + }, + { + "epoch": 1.333256110274528, + "grad_norm": 1.7840913534164429, + "learning_rate": 0.00011111883341699166, + "loss": 0.1759, + "step": 34530 + }, + { + "epoch": 1.3336422255685547, + "grad_norm": 1.1825993061065674, + "learning_rate": 0.00011109309239738987, + "loss": 0.2464, + "step": 34540 + }, + { + "epoch": 1.3340283408625815, + "grad_norm": 1.8859659433364868, + "learning_rate": 0.00011106735137778807, + "loss": 0.3539, + "step": 34550 + }, + { + "epoch": 1.3344144561566083, + "grad_norm": 1.9698175191879272, + "learning_rate": 0.0001110416103581863, + "loss": 0.3301, + "step": 34560 + }, + { + "epoch": 1.3348005714506352, + "grad_norm": 0.7649385333061218, + "learning_rate": 0.0001110158693385845, + "loss": 0.232, + "step": 34570 + }, + { + "epoch": 1.335186686744662, + "grad_norm": 0.56386399269104, + "learning_rate": 0.00011099012831898272, + "loss": 0.3425, + "step": 34580 + }, + { + "epoch": 1.3355728020386888, + "grad_norm": 2.956003189086914, + "learning_rate": 0.00011096438729938092, + "loss": 0.1518, + "step": 34590 + }, + { + "epoch": 1.3359589173327155, + "grad_norm": 2.612029552459717, + "learning_rate": 0.00011093864627977915, + "loss": 0.2765, + "step": 34600 + }, + { + "epoch": 1.3363450326267423, + "grad_norm": 0.9674397706985474, + "learning_rate": 0.00011091290526017736, + "loss": 0.303, + "step": 34610 + }, + { + "epoch": 1.336731147920769, + "grad_norm": 0.9578921794891357, + "learning_rate": 0.00011088716424057556, + "loss": 0.1405, + "step": 34620 + }, + { + "epoch": 1.3371172632147958, + "grad_norm": 2.168065071105957, + "learning_rate": 0.00011086142322097379, + "loss": 0.1914, + "step": 34630 + }, + { + "epoch": 1.3375033785088228, + "grad_norm": 1.3166526556015015, + "learning_rate": 0.00011083568220137199, + "loss": 0.4134, + "step": 34640 + }, + { + "epoch": 1.3378894938028496, + "grad_norm": 0.9082283973693848, + "learning_rate": 0.00011080994118177022, + "loss": 0.2693, + "step": 34650 + }, + { + "epoch": 1.3382756090968764, + "grad_norm": 2.203007698059082, + "learning_rate": 0.00011078420016216844, + "loss": 0.1847, + "step": 34660 + }, + { + "epoch": 1.3386617243909031, + "grad_norm": 0.8101674914360046, + "learning_rate": 0.00011075845914256664, + "loss": 0.3111, + "step": 34670 + }, + { + "epoch": 1.33904783968493, + "grad_norm": 1.9545695781707764, + "learning_rate": 0.00011073271812296486, + "loss": 0.3761, + "step": 34680 + }, + { + "epoch": 1.3394339549789567, + "grad_norm": 1.547581672668457, + "learning_rate": 0.00011070697710336306, + "loss": 0.2374, + "step": 34690 + }, + { + "epoch": 1.3398200702729834, + "grad_norm": 3.3519034385681152, + "learning_rate": 0.00011068123608376128, + "loss": 0.1957, + "step": 34700 + }, + { + "epoch": 1.3402061855670104, + "grad_norm": 1.5508599281311035, + "learning_rate": 0.00011065549506415948, + "loss": 0.4171, + "step": 34710 + }, + { + "epoch": 1.340592300861037, + "grad_norm": 1.8547546863555908, + "learning_rate": 0.00011062975404455771, + "loss": 0.1872, + "step": 34720 + }, + { + "epoch": 1.340978416155064, + "grad_norm": 1.4600756168365479, + "learning_rate": 0.00011060401302495594, + "loss": 0.3515, + "step": 34730 + }, + { + "epoch": 1.3413645314490907, + "grad_norm": 0.05774044618010521, + "learning_rate": 0.00011057827200535414, + "loss": 0.1604, + "step": 34740 + }, + { + "epoch": 1.3417506467431175, + "grad_norm": 2.8793342113494873, + "learning_rate": 0.00011055253098575235, + "loss": 0.3095, + "step": 34750 + }, + { + "epoch": 1.3421367620371443, + "grad_norm": 2.241042375564575, + "learning_rate": 0.00011052678996615055, + "loss": 0.2511, + "step": 34760 + }, + { + "epoch": 1.342522877331171, + "grad_norm": 1.9320632219314575, + "learning_rate": 0.00011050104894654878, + "loss": 0.4493, + "step": 34770 + }, + { + "epoch": 1.3429089926251978, + "grad_norm": 1.6483882665634155, + "learning_rate": 0.000110475307926947, + "loss": 0.217, + "step": 34780 + }, + { + "epoch": 1.3432951079192246, + "grad_norm": 0.9635765552520752, + "learning_rate": 0.0001104495669073452, + "loss": 0.5458, + "step": 34790 + }, + { + "epoch": 1.3436812232132516, + "grad_norm": 1.2436567544937134, + "learning_rate": 0.00011042382588774343, + "loss": 0.2857, + "step": 34800 + }, + { + "epoch": 1.3440673385072783, + "grad_norm": 2.8082425594329834, + "learning_rate": 0.00011039808486814163, + "loss": 0.3439, + "step": 34810 + }, + { + "epoch": 1.344453453801305, + "grad_norm": 1.0430901050567627, + "learning_rate": 0.00011037234384853984, + "loss": 0.1404, + "step": 34820 + }, + { + "epoch": 1.3448395690953319, + "grad_norm": 1.7387149333953857, + "learning_rate": 0.00011034660282893806, + "loss": 0.395, + "step": 34830 + }, + { + "epoch": 1.3452256843893586, + "grad_norm": 1.2713748216629028, + "learning_rate": 0.00011032086180933627, + "loss": 0.29, + "step": 34840 + }, + { + "epoch": 1.3456117996833854, + "grad_norm": 0.26068204641342163, + "learning_rate": 0.0001102951207897345, + "loss": 0.1814, + "step": 34850 + }, + { + "epoch": 1.3459979149774122, + "grad_norm": 2.4163243770599365, + "learning_rate": 0.0001102693797701327, + "loss": 0.2445, + "step": 34860 + }, + { + "epoch": 1.3463840302714392, + "grad_norm": 2.2439687252044678, + "learning_rate": 0.00011024363875053092, + "loss": 0.3338, + "step": 34870 + }, + { + "epoch": 1.346770145565466, + "grad_norm": 0.2822403609752655, + "learning_rate": 0.00011021789773092912, + "loss": 0.1648, + "step": 34880 + }, + { + "epoch": 1.3471562608594927, + "grad_norm": 0.07319017499685287, + "learning_rate": 0.00011019215671132734, + "loss": 0.107, + "step": 34890 + }, + { + "epoch": 1.3475423761535195, + "grad_norm": 0.9809044003486633, + "learning_rate": 0.00011016641569172555, + "loss": 0.256, + "step": 34900 + }, + { + "epoch": 1.3479284914475462, + "grad_norm": 0.5016226768493652, + "learning_rate": 0.00011014067467212376, + "loss": 0.3229, + "step": 34910 + }, + { + "epoch": 1.348314606741573, + "grad_norm": 1.3026005029678345, + "learning_rate": 0.00011011493365252199, + "loss": 0.2145, + "step": 34920 + }, + { + "epoch": 1.3487007220355998, + "grad_norm": 1.0752215385437012, + "learning_rate": 0.00011008919263292019, + "loss": 0.2355, + "step": 34930 + }, + { + "epoch": 1.3490868373296268, + "grad_norm": 2.2703003883361816, + "learning_rate": 0.00011006345161331842, + "loss": 0.2079, + "step": 34940 + }, + { + "epoch": 1.3494729526236533, + "grad_norm": 1.1323810815811157, + "learning_rate": 0.00011003771059371662, + "loss": 0.1015, + "step": 34950 + }, + { + "epoch": 1.3498590679176803, + "grad_norm": 0.10813555121421814, + "learning_rate": 0.00011001196957411484, + "loss": 0.4214, + "step": 34960 + }, + { + "epoch": 1.350245183211707, + "grad_norm": 0.07815568149089813, + "learning_rate": 0.00010998622855451306, + "loss": 0.1211, + "step": 34970 + }, + { + "epoch": 1.3506312985057338, + "grad_norm": 0.6748234629631042, + "learning_rate": 0.00010996048753491126, + "loss": 0.3508, + "step": 34980 + }, + { + "epoch": 1.3510174137997606, + "grad_norm": 1.8556997776031494, + "learning_rate": 0.00010993474651530948, + "loss": 0.2268, + "step": 34990 + }, + { + "epoch": 1.3514035290937874, + "grad_norm": 0.8696061372756958, + "learning_rate": 0.00010990900549570768, + "loss": 0.4321, + "step": 35000 + }, + { + "epoch": 1.3517896443878141, + "grad_norm": 0.42442765831947327, + "learning_rate": 0.00010988326447610591, + "loss": 0.1944, + "step": 35010 + }, + { + "epoch": 1.352175759681841, + "grad_norm": 1.0474554300308228, + "learning_rate": 0.00010985752345650411, + "loss": 0.1342, + "step": 35020 + }, + { + "epoch": 1.3525618749758679, + "grad_norm": 0.607037365436554, + "learning_rate": 0.00010983178243690234, + "loss": 0.2965, + "step": 35030 + }, + { + "epoch": 1.3529479902698947, + "grad_norm": 1.8160990476608276, + "learning_rate": 0.00010980604141730055, + "loss": 0.3192, + "step": 35040 + }, + { + "epoch": 1.3533341055639214, + "grad_norm": 2.0026509761810303, + "learning_rate": 0.00010978030039769875, + "loss": 0.3054, + "step": 35050 + }, + { + "epoch": 1.3537202208579482, + "grad_norm": 0.9203600883483887, + "learning_rate": 0.00010975455937809698, + "loss": 0.253, + "step": 35060 + }, + { + "epoch": 1.354106336151975, + "grad_norm": 0.33198195695877075, + "learning_rate": 0.00010972881835849518, + "loss": 0.3885, + "step": 35070 + }, + { + "epoch": 1.3544924514460017, + "grad_norm": 0.3201223611831665, + "learning_rate": 0.0001097030773388934, + "loss": 0.3029, + "step": 35080 + }, + { + "epoch": 1.3548785667400285, + "grad_norm": 1.2589943408966064, + "learning_rate": 0.0001096773363192916, + "loss": 0.4243, + "step": 35090 + }, + { + "epoch": 1.3552646820340555, + "grad_norm": 1.5106219053268433, + "learning_rate": 0.00010965159529968983, + "loss": 0.2585, + "step": 35100 + }, + { + "epoch": 1.3556507973280822, + "grad_norm": 1.429799199104309, + "learning_rate": 0.00010962585428008804, + "loss": 0.1961, + "step": 35110 + }, + { + "epoch": 1.356036912622109, + "grad_norm": 2.1211297512054443, + "learning_rate": 0.00010960011326048624, + "loss": 0.4057, + "step": 35120 + }, + { + "epoch": 1.3564230279161358, + "grad_norm": 2.5154731273651123, + "learning_rate": 0.00010957437224088447, + "loss": 0.3787, + "step": 35130 + }, + { + "epoch": 1.3568091432101625, + "grad_norm": 0.4914834201335907, + "learning_rate": 0.00010954863122128267, + "loss": 0.234, + "step": 35140 + }, + { + "epoch": 1.3571952585041893, + "grad_norm": 0.26685893535614014, + "learning_rate": 0.0001095228902016809, + "loss": 0.2841, + "step": 35150 + }, + { + "epoch": 1.357581373798216, + "grad_norm": 0.15462155640125275, + "learning_rate": 0.00010949714918207912, + "loss": 0.2269, + "step": 35160 + }, + { + "epoch": 1.357967489092243, + "grad_norm": 1.3887063264846802, + "learning_rate": 0.00010947140816247732, + "loss": 0.3455, + "step": 35170 + }, + { + "epoch": 1.3583536043862696, + "grad_norm": 0.786374032497406, + "learning_rate": 0.00010944566714287554, + "loss": 0.2897, + "step": 35180 + }, + { + "epoch": 1.3587397196802966, + "grad_norm": 1.100475549697876, + "learning_rate": 0.00010941992612327374, + "loss": 0.2892, + "step": 35190 + }, + { + "epoch": 1.3591258349743234, + "grad_norm": 0.7676102519035339, + "learning_rate": 0.00010939418510367196, + "loss": 0.1942, + "step": 35200 + }, + { + "epoch": 1.3595119502683501, + "grad_norm": 0.33462053537368774, + "learning_rate": 0.00010936844408407016, + "loss": 0.2872, + "step": 35210 + }, + { + "epoch": 1.359898065562377, + "grad_norm": 0.9294387698173523, + "learning_rate": 0.00010934270306446839, + "loss": 0.2617, + "step": 35220 + }, + { + "epoch": 1.3602841808564037, + "grad_norm": 0.3169979453086853, + "learning_rate": 0.00010931696204486662, + "loss": 0.2942, + "step": 35230 + }, + { + "epoch": 1.3606702961504304, + "grad_norm": 2.1339616775512695, + "learning_rate": 0.00010929122102526482, + "loss": 0.4448, + "step": 35240 + }, + { + "epoch": 1.3610564114444572, + "grad_norm": 0.9430062770843506, + "learning_rate": 0.00010926548000566303, + "loss": 0.2051, + "step": 35250 + }, + { + "epoch": 1.3614425267384842, + "grad_norm": 3.1187360286712646, + "learning_rate": 0.00010923973898606123, + "loss": 0.2274, + "step": 35260 + }, + { + "epoch": 1.361828642032511, + "grad_norm": 1.4727579355239868, + "learning_rate": 0.00010921399796645946, + "loss": 0.3757, + "step": 35270 + }, + { + "epoch": 1.3622147573265377, + "grad_norm": 2.157560348510742, + "learning_rate": 0.00010918825694685768, + "loss": 0.3096, + "step": 35280 + }, + { + "epoch": 1.3626008726205645, + "grad_norm": 0.33457377552986145, + "learning_rate": 0.00010916251592725588, + "loss": 0.1489, + "step": 35290 + }, + { + "epoch": 1.3629869879145913, + "grad_norm": 0.9005904197692871, + "learning_rate": 0.00010913677490765411, + "loss": 0.1826, + "step": 35300 + }, + { + "epoch": 1.363373103208618, + "grad_norm": 2.1222829818725586, + "learning_rate": 0.00010911103388805231, + "loss": 0.1965, + "step": 35310 + }, + { + "epoch": 1.3637592185026448, + "grad_norm": 1.3881357908248901, + "learning_rate": 0.00010908529286845052, + "loss": 0.1791, + "step": 35320 + }, + { + "epoch": 1.3641453337966718, + "grad_norm": 1.7574503421783447, + "learning_rate": 0.00010905955184884872, + "loss": 0.3316, + "step": 35330 + }, + { + "epoch": 1.3645314490906983, + "grad_norm": 0.1967727392911911, + "learning_rate": 0.00010903381082924695, + "loss": 0.2331, + "step": 35340 + }, + { + "epoch": 1.3649175643847253, + "grad_norm": 0.8974360823631287, + "learning_rate": 0.00010900806980964518, + "loss": 0.2589, + "step": 35350 + }, + { + "epoch": 1.365303679678752, + "grad_norm": 2.0996744632720947, + "learning_rate": 0.00010898232879004338, + "loss": 0.3663, + "step": 35360 + }, + { + "epoch": 1.3656897949727789, + "grad_norm": 0.5678316354751587, + "learning_rate": 0.0001089565877704416, + "loss": 0.1729, + "step": 35370 + }, + { + "epoch": 1.3660759102668056, + "grad_norm": 2.3381874561309814, + "learning_rate": 0.0001089308467508398, + "loss": 0.1615, + "step": 35380 + }, + { + "epoch": 1.3664620255608324, + "grad_norm": 1.0276836156845093, + "learning_rate": 0.00010890510573123802, + "loss": 0.3359, + "step": 35390 + }, + { + "epoch": 1.3668481408548594, + "grad_norm": 2.4374940395355225, + "learning_rate": 0.00010887936471163622, + "loss": 0.2435, + "step": 35400 + }, + { + "epoch": 1.367234256148886, + "grad_norm": 0.45221665501594543, + "learning_rate": 0.00010885362369203444, + "loss": 0.2555, + "step": 35410 + }, + { + "epoch": 1.367620371442913, + "grad_norm": 2.608090400695801, + "learning_rate": 0.00010882788267243267, + "loss": 0.2465, + "step": 35420 + }, + { + "epoch": 1.3680064867369397, + "grad_norm": 1.4186642169952393, + "learning_rate": 0.00010880214165283087, + "loss": 0.1674, + "step": 35430 + }, + { + "epoch": 1.3683926020309665, + "grad_norm": 0.659479022026062, + "learning_rate": 0.0001087764006332291, + "loss": 0.2926, + "step": 35440 + }, + { + "epoch": 1.3687787173249932, + "grad_norm": 0.9219567179679871, + "learning_rate": 0.0001087506596136273, + "loss": 0.2001, + "step": 35450 + }, + { + "epoch": 1.36916483261902, + "grad_norm": 0.8070804476737976, + "learning_rate": 0.00010872491859402551, + "loss": 0.2178, + "step": 35460 + }, + { + "epoch": 1.3695509479130468, + "grad_norm": 2.9981069564819336, + "learning_rate": 0.00010869917757442374, + "loss": 0.3079, + "step": 35470 + }, + { + "epoch": 1.3699370632070735, + "grad_norm": 0.7891242504119873, + "learning_rate": 0.00010867343655482194, + "loss": 0.2765, + "step": 35480 + }, + { + "epoch": 1.3703231785011005, + "grad_norm": 1.448637843132019, + "learning_rate": 0.00010864769553522016, + "loss": 0.3521, + "step": 35490 + }, + { + "epoch": 1.3707092937951273, + "grad_norm": 0.07628043740987778, + "learning_rate": 0.00010862195451561836, + "loss": 0.2083, + "step": 35500 + }, + { + "epoch": 1.371095409089154, + "grad_norm": 0.7549735307693481, + "learning_rate": 0.00010859621349601659, + "loss": 0.2536, + "step": 35510 + }, + { + "epoch": 1.3714815243831808, + "grad_norm": 1.3548041582107544, + "learning_rate": 0.00010857047247641479, + "loss": 0.251, + "step": 35520 + }, + { + "epoch": 1.3718676396772076, + "grad_norm": 0.530010998249054, + "learning_rate": 0.000108544731456813, + "loss": 0.1917, + "step": 35530 + }, + { + "epoch": 1.3722537549712344, + "grad_norm": 0.4148992896080017, + "learning_rate": 0.00010851899043721123, + "loss": 0.335, + "step": 35540 + }, + { + "epoch": 1.3726398702652611, + "grad_norm": 1.5118776559829712, + "learning_rate": 0.00010849324941760943, + "loss": 0.2159, + "step": 35550 + }, + { + "epoch": 1.3730259855592881, + "grad_norm": 1.036889910697937, + "learning_rate": 0.00010846750839800766, + "loss": 0.2975, + "step": 35560 + }, + { + "epoch": 1.3734121008533147, + "grad_norm": 1.724263072013855, + "learning_rate": 0.00010844176737840586, + "loss": 0.1476, + "step": 35570 + }, + { + "epoch": 1.3737982161473417, + "grad_norm": 1.599007487297058, + "learning_rate": 0.00010841602635880408, + "loss": 0.2539, + "step": 35580 + }, + { + "epoch": 1.3741843314413684, + "grad_norm": 2.9119279384613037, + "learning_rate": 0.00010839028533920228, + "loss": 0.2688, + "step": 35590 + }, + { + "epoch": 1.3745704467353952, + "grad_norm": 1.8647874593734741, + "learning_rate": 0.00010836454431960051, + "loss": 0.4158, + "step": 35600 + }, + { + "epoch": 1.374956562029422, + "grad_norm": 3.925290822982788, + "learning_rate": 0.00010833880329999872, + "loss": 0.3333, + "step": 35610 + }, + { + "epoch": 1.3753426773234487, + "grad_norm": 0.7124634385108948, + "learning_rate": 0.00010831306228039692, + "loss": 0.1069, + "step": 35620 + }, + { + "epoch": 1.3757287926174757, + "grad_norm": 1.303579330444336, + "learning_rate": 0.00010828732126079515, + "loss": 0.2898, + "step": 35630 + }, + { + "epoch": 1.3761149079115023, + "grad_norm": 3.921804189682007, + "learning_rate": 0.00010826158024119335, + "loss": 0.4212, + "step": 35640 + }, + { + "epoch": 1.3765010232055293, + "grad_norm": 1.3194564580917358, + "learning_rate": 0.00010823583922159158, + "loss": 0.2771, + "step": 35650 + }, + { + "epoch": 1.376887138499556, + "grad_norm": 1.4237637519836426, + "learning_rate": 0.00010821009820198979, + "loss": 0.2463, + "step": 35660 + }, + { + "epoch": 1.3772732537935828, + "grad_norm": 1.8165888786315918, + "learning_rate": 0.000108184357182388, + "loss": 0.291, + "step": 35670 + }, + { + "epoch": 1.3776593690876096, + "grad_norm": 1.1056426763534546, + "learning_rate": 0.00010815861616278622, + "loss": 0.2525, + "step": 35680 + }, + { + "epoch": 1.3780454843816363, + "grad_norm": 1.483189582824707, + "learning_rate": 0.00010813287514318442, + "loss": 0.1569, + "step": 35690 + }, + { + "epoch": 1.378431599675663, + "grad_norm": 1.0666841268539429, + "learning_rate": 0.00010810713412358264, + "loss": 0.235, + "step": 35700 + }, + { + "epoch": 1.3788177149696899, + "grad_norm": 1.0299845933914185, + "learning_rate": 0.00010808139310398084, + "loss": 0.3892, + "step": 35710 + }, + { + "epoch": 1.3792038302637168, + "grad_norm": 2.3474409580230713, + "learning_rate": 0.00010805565208437907, + "loss": 0.3417, + "step": 35720 + }, + { + "epoch": 1.3795899455577436, + "grad_norm": 1.7456315755844116, + "learning_rate": 0.0001080299110647773, + "loss": 0.2538, + "step": 35730 + }, + { + "epoch": 1.3799760608517704, + "grad_norm": 2.866103410720825, + "learning_rate": 0.0001080041700451755, + "loss": 0.1619, + "step": 35740 + }, + { + "epoch": 1.3803621761457971, + "grad_norm": 0.29136407375335693, + "learning_rate": 0.00010797842902557371, + "loss": 0.2692, + "step": 35750 + }, + { + "epoch": 1.380748291439824, + "grad_norm": 0.8046161532402039, + "learning_rate": 0.00010795268800597191, + "loss": 0.1575, + "step": 35760 + }, + { + "epoch": 1.3811344067338507, + "grad_norm": 0.6451787352561951, + "learning_rate": 0.00010792694698637014, + "loss": 0.4914, + "step": 35770 + }, + { + "epoch": 1.3815205220278775, + "grad_norm": 0.7289161086082458, + "learning_rate": 0.00010790120596676836, + "loss": 0.1895, + "step": 35780 + }, + { + "epoch": 1.3819066373219044, + "grad_norm": 0.8300430178642273, + "learning_rate": 0.00010787546494716656, + "loss": 0.3663, + "step": 35790 + }, + { + "epoch": 1.382292752615931, + "grad_norm": 0.17713364958763123, + "learning_rate": 0.00010784972392756479, + "loss": 0.3189, + "step": 35800 + }, + { + "epoch": 1.382678867909958, + "grad_norm": 0.903222918510437, + "learning_rate": 0.00010782398290796299, + "loss": 0.1577, + "step": 35810 + }, + { + "epoch": 1.3830649832039847, + "grad_norm": 0.08617932349443436, + "learning_rate": 0.0001077982418883612, + "loss": 0.2872, + "step": 35820 + }, + { + "epoch": 1.3834510984980115, + "grad_norm": 1.9590895175933838, + "learning_rate": 0.0001077725008687594, + "loss": 0.2907, + "step": 35830 + }, + { + "epoch": 1.3838372137920383, + "grad_norm": 1.2515161037445068, + "learning_rate": 0.00010774675984915763, + "loss": 0.177, + "step": 35840 + }, + { + "epoch": 1.384223329086065, + "grad_norm": 1.6171292066574097, + "learning_rate": 0.00010772101882955586, + "loss": 0.2321, + "step": 35850 + }, + { + "epoch": 1.3846094443800918, + "grad_norm": 0.13681405782699585, + "learning_rate": 0.00010769527780995406, + "loss": 0.247, + "step": 35860 + }, + { + "epoch": 1.3849955596741186, + "grad_norm": 1.1949968338012695, + "learning_rate": 0.00010766953679035228, + "loss": 0.2978, + "step": 35870 + }, + { + "epoch": 1.3853816749681456, + "grad_norm": 0.17001692950725555, + "learning_rate": 0.00010764379577075048, + "loss": 0.3241, + "step": 35880 + }, + { + "epoch": 1.3857677902621723, + "grad_norm": 0.8227952122688293, + "learning_rate": 0.0001076180547511487, + "loss": 0.3499, + "step": 35890 + }, + { + "epoch": 1.386153905556199, + "grad_norm": 1.4185482263565063, + "learning_rate": 0.0001075923137315469, + "loss": 0.3109, + "step": 35900 + }, + { + "epoch": 1.3865400208502259, + "grad_norm": 0.9533351063728333, + "learning_rate": 0.00010756657271194512, + "loss": 0.24, + "step": 35910 + }, + { + "epoch": 1.3869261361442526, + "grad_norm": 1.077789306640625, + "learning_rate": 0.00010754083169234335, + "loss": 0.2662, + "step": 35920 + }, + { + "epoch": 1.3873122514382794, + "grad_norm": 1.3528363704681396, + "learning_rate": 0.00010751509067274155, + "loss": 0.1623, + "step": 35930 + }, + { + "epoch": 1.3876983667323062, + "grad_norm": 0.25122806429862976, + "learning_rate": 0.00010748934965313978, + "loss": 0.1639, + "step": 35940 + }, + { + "epoch": 1.3880844820263332, + "grad_norm": 0.9446159601211548, + "learning_rate": 0.00010746360863353798, + "loss": 0.2035, + "step": 35950 + }, + { + "epoch": 1.38847059732036, + "grad_norm": 1.2258719205856323, + "learning_rate": 0.00010743786761393619, + "loss": 0.4247, + "step": 35960 + }, + { + "epoch": 1.3888567126143867, + "grad_norm": 3.117729663848877, + "learning_rate": 0.00010741212659433442, + "loss": 0.2569, + "step": 35970 + }, + { + "epoch": 1.3892428279084135, + "grad_norm": 0.6333123445510864, + "learning_rate": 0.00010738638557473262, + "loss": 0.1843, + "step": 35980 + }, + { + "epoch": 1.3896289432024402, + "grad_norm": 1.49360191822052, + "learning_rate": 0.00010736064455513084, + "loss": 0.2318, + "step": 35990 + }, + { + "epoch": 1.390015058496467, + "grad_norm": 3.9082753658294678, + "learning_rate": 0.00010733490353552904, + "loss": 0.2037, + "step": 36000 + }, + { + "epoch": 1.3904011737904938, + "grad_norm": 0.5687323808670044, + "learning_rate": 0.00010730916251592727, + "loss": 0.0871, + "step": 36010 + }, + { + "epoch": 1.3907872890845208, + "grad_norm": 0.3480868339538574, + "learning_rate": 0.00010728342149632547, + "loss": 0.4009, + "step": 36020 + }, + { + "epoch": 1.3911734043785473, + "grad_norm": 1.325042486190796, + "learning_rate": 0.00010725768047672368, + "loss": 0.5143, + "step": 36030 + }, + { + "epoch": 1.3915595196725743, + "grad_norm": 2.114786386489868, + "learning_rate": 0.00010723193945712191, + "loss": 0.2944, + "step": 36040 + }, + { + "epoch": 1.391945634966601, + "grad_norm": 1.716272234916687, + "learning_rate": 0.00010720619843752011, + "loss": 0.3793, + "step": 36050 + }, + { + "epoch": 1.3923317502606278, + "grad_norm": 2.057535171508789, + "learning_rate": 0.00010718045741791834, + "loss": 0.199, + "step": 36060 + }, + { + "epoch": 1.3927178655546546, + "grad_norm": 2.371248483657837, + "learning_rate": 0.00010715471639831653, + "loss": 0.3276, + "step": 36070 + }, + { + "epoch": 1.3931039808486814, + "grad_norm": 0.1784186065196991, + "learning_rate": 0.00010712897537871476, + "loss": 0.0922, + "step": 36080 + }, + { + "epoch": 1.3934900961427081, + "grad_norm": 0.7923040390014648, + "learning_rate": 0.00010710323435911298, + "loss": 0.1599, + "step": 36090 + }, + { + "epoch": 1.393876211436735, + "grad_norm": 1.9013831615447998, + "learning_rate": 0.00010707749333951117, + "loss": 0.4997, + "step": 36100 + }, + { + "epoch": 1.394262326730762, + "grad_norm": 1.8659415245056152, + "learning_rate": 0.0001070517523199094, + "loss": 0.1851, + "step": 36110 + }, + { + "epoch": 1.3946484420247887, + "grad_norm": 2.5775375366210938, + "learning_rate": 0.0001070260113003076, + "loss": 0.2384, + "step": 36120 + }, + { + "epoch": 1.3950345573188154, + "grad_norm": 0.21943879127502441, + "learning_rate": 0.00010700027028070583, + "loss": 0.3996, + "step": 36130 + }, + { + "epoch": 1.3954206726128422, + "grad_norm": 1.1734743118286133, + "learning_rate": 0.00010697452926110403, + "loss": 0.1737, + "step": 36140 + }, + { + "epoch": 1.395806787906869, + "grad_norm": 0.697695791721344, + "learning_rate": 0.00010694878824150225, + "loss": 0.2778, + "step": 36150 + }, + { + "epoch": 1.3961929032008957, + "grad_norm": 3.2881579399108887, + "learning_rate": 0.00010692304722190047, + "loss": 0.3198, + "step": 36160 + }, + { + "epoch": 1.3965790184949225, + "grad_norm": 0.1592467725276947, + "learning_rate": 0.00010689730620229868, + "loss": 0.2591, + "step": 36170 + }, + { + "epoch": 1.3969651337889495, + "grad_norm": 0.4579029083251953, + "learning_rate": 0.0001068715651826969, + "loss": 0.2984, + "step": 36180 + }, + { + "epoch": 1.3973512490829763, + "grad_norm": 0.4286015033721924, + "learning_rate": 0.0001068458241630951, + "loss": 0.4098, + "step": 36190 + }, + { + "epoch": 1.397737364377003, + "grad_norm": 1.7824127674102783, + "learning_rate": 0.00010682008314349332, + "loss": 0.2446, + "step": 36200 + }, + { + "epoch": 1.3981234796710298, + "grad_norm": 0.8584449887275696, + "learning_rate": 0.00010679434212389152, + "loss": 0.1415, + "step": 36210 + }, + { + "epoch": 1.3985095949650566, + "grad_norm": 1.1699339151382446, + "learning_rate": 0.00010676860110428975, + "loss": 0.1706, + "step": 36220 + }, + { + "epoch": 1.3988957102590833, + "grad_norm": 2.615877389907837, + "learning_rate": 0.00010674286008468796, + "loss": 0.2605, + "step": 36230 + }, + { + "epoch": 1.39928182555311, + "grad_norm": 2.182037591934204, + "learning_rate": 0.00010671711906508617, + "loss": 0.2067, + "step": 36240 + }, + { + "epoch": 1.399667940847137, + "grad_norm": 2.183263063430786, + "learning_rate": 0.00010669137804548439, + "loss": 0.4704, + "step": 36250 + }, + { + "epoch": 1.4000540561411636, + "grad_norm": 3.505791187286377, + "learning_rate": 0.00010666563702588259, + "loss": 0.3385, + "step": 36260 + }, + { + "epoch": 1.4004401714351906, + "grad_norm": 1.2262030839920044, + "learning_rate": 0.00010663989600628081, + "loss": 0.3999, + "step": 36270 + }, + { + "epoch": 1.4008262867292174, + "grad_norm": 2.4024577140808105, + "learning_rate": 0.00010661415498667904, + "loss": 0.1984, + "step": 36280 + }, + { + "epoch": 1.4012124020232442, + "grad_norm": 0.4166090786457062, + "learning_rate": 0.00010658841396707724, + "loss": 0.2448, + "step": 36290 + }, + { + "epoch": 1.401598517317271, + "grad_norm": 0.422590047121048, + "learning_rate": 0.00010656267294747547, + "loss": 0.2464, + "step": 36300 + }, + { + "epoch": 1.4019846326112977, + "grad_norm": 2.287503480911255, + "learning_rate": 0.00010653693192787367, + "loss": 0.1621, + "step": 36310 + }, + { + "epoch": 1.4023707479053245, + "grad_norm": 0.8126110434532166, + "learning_rate": 0.00010651119090827188, + "loss": 0.2097, + "step": 36320 + }, + { + "epoch": 1.4027568631993512, + "grad_norm": 0.683016836643219, + "learning_rate": 0.00010648544988867008, + "loss": 0.1512, + "step": 36330 + }, + { + "epoch": 1.4031429784933782, + "grad_norm": 1.5477893352508545, + "learning_rate": 0.00010645970886906831, + "loss": 0.2169, + "step": 36340 + }, + { + "epoch": 1.403529093787405, + "grad_norm": 2.183166265487671, + "learning_rate": 0.00010643396784946653, + "loss": 0.4307, + "step": 36350 + }, + { + "epoch": 1.4039152090814317, + "grad_norm": 1.5782747268676758, + "learning_rate": 0.00010640822682986473, + "loss": 0.291, + "step": 36360 + }, + { + "epoch": 1.4043013243754585, + "grad_norm": 1.2823392152786255, + "learning_rate": 0.00010638248581026296, + "loss": 0.315, + "step": 36370 + }, + { + "epoch": 1.4046874396694853, + "grad_norm": 2.0126500129699707, + "learning_rate": 0.00010635674479066116, + "loss": 0.2565, + "step": 36380 + }, + { + "epoch": 1.405073554963512, + "grad_norm": 2.490217447280884, + "learning_rate": 0.00010633100377105937, + "loss": 0.2665, + "step": 36390 + }, + { + "epoch": 1.4054596702575388, + "grad_norm": 1.2054855823516846, + "learning_rate": 0.00010630526275145757, + "loss": 0.5455, + "step": 36400 + }, + { + "epoch": 1.4058457855515658, + "grad_norm": 0.7968757748603821, + "learning_rate": 0.0001062795217318558, + "loss": 0.333, + "step": 36410 + }, + { + "epoch": 1.4062319008455926, + "grad_norm": 1.1027718782424927, + "learning_rate": 0.00010625378071225403, + "loss": 0.2019, + "step": 36420 + }, + { + "epoch": 1.4066180161396193, + "grad_norm": 1.987302541732788, + "learning_rate": 0.00010622803969265223, + "loss": 0.3159, + "step": 36430 + }, + { + "epoch": 1.407004131433646, + "grad_norm": 1.5426512956619263, + "learning_rate": 0.00010620229867305045, + "loss": 0.3759, + "step": 36440 + }, + { + "epoch": 1.4073902467276729, + "grad_norm": 0.5228156447410583, + "learning_rate": 0.00010617655765344865, + "loss": 0.2465, + "step": 36450 + }, + { + "epoch": 1.4077763620216996, + "grad_norm": 0.46890121698379517, + "learning_rate": 0.00010615081663384687, + "loss": 0.302, + "step": 36460 + }, + { + "epoch": 1.4081624773157264, + "grad_norm": 2.1506495475769043, + "learning_rate": 0.0001061250756142451, + "loss": 0.2569, + "step": 36470 + }, + { + "epoch": 1.4085485926097534, + "grad_norm": 2.307468891143799, + "learning_rate": 0.0001060993345946433, + "loss": 0.2009, + "step": 36480 + }, + { + "epoch": 1.40893470790378, + "grad_norm": 0.07033026963472366, + "learning_rate": 0.00010607359357504152, + "loss": 0.1728, + "step": 36490 + }, + { + "epoch": 1.409320823197807, + "grad_norm": 0.3262972831726074, + "learning_rate": 0.00010604785255543972, + "loss": 0.2905, + "step": 36500 + }, + { + "epoch": 1.4097069384918337, + "grad_norm": 0.755646824836731, + "learning_rate": 0.00010602211153583795, + "loss": 0.3287, + "step": 36510 + }, + { + "epoch": 1.4100930537858605, + "grad_norm": 1.1180161237716675, + "learning_rate": 0.00010599637051623615, + "loss": 0.2743, + "step": 36520 + }, + { + "epoch": 1.4104791690798872, + "grad_norm": 1.2358392477035522, + "learning_rate": 0.00010597062949663436, + "loss": 0.2672, + "step": 36530 + }, + { + "epoch": 1.410865284373914, + "grad_norm": 1.4188988208770752, + "learning_rate": 0.00010594488847703259, + "loss": 0.2552, + "step": 36540 + }, + { + "epoch": 1.4112513996679408, + "grad_norm": 1.4727978706359863, + "learning_rate": 0.00010591914745743079, + "loss": 0.2256, + "step": 36550 + }, + { + "epoch": 1.4116375149619675, + "grad_norm": 0.08973213285207748, + "learning_rate": 0.00010589340643782901, + "loss": 0.224, + "step": 36560 + }, + { + "epoch": 1.4120236302559945, + "grad_norm": 0.9915102124214172, + "learning_rate": 0.00010586766541822721, + "loss": 0.195, + "step": 36570 + }, + { + "epoch": 1.4124097455500213, + "grad_norm": 0.8524800539016724, + "learning_rate": 0.00010584192439862544, + "loss": 0.1492, + "step": 36580 + }, + { + "epoch": 1.412795860844048, + "grad_norm": 1.8414466381072998, + "learning_rate": 0.00010581618337902365, + "loss": 0.3412, + "step": 36590 + }, + { + "epoch": 1.4131819761380748, + "grad_norm": 2.596547842025757, + "learning_rate": 0.00010579044235942185, + "loss": 0.3469, + "step": 36600 + }, + { + "epoch": 1.4135680914321016, + "grad_norm": 0.5367813110351562, + "learning_rate": 0.00010576470133982008, + "loss": 0.2274, + "step": 36610 + }, + { + "epoch": 1.4139542067261284, + "grad_norm": 2.70858097076416, + "learning_rate": 0.00010573896032021828, + "loss": 0.171, + "step": 36620 + }, + { + "epoch": 1.4143403220201551, + "grad_norm": 1.7077667713165283, + "learning_rate": 0.00010571321930061651, + "loss": 0.2638, + "step": 36630 + }, + { + "epoch": 1.4147264373141821, + "grad_norm": 0.9189953804016113, + "learning_rate": 0.00010568747828101471, + "loss": 0.2283, + "step": 36640 + }, + { + "epoch": 1.4151125526082087, + "grad_norm": 2.0240087509155273, + "learning_rate": 0.00010566173726141293, + "loss": 0.3166, + "step": 36650 + }, + { + "epoch": 1.4154986679022357, + "grad_norm": 3.3304507732391357, + "learning_rate": 0.00010563599624181115, + "loss": 0.3046, + "step": 36660 + }, + { + "epoch": 1.4158847831962624, + "grad_norm": 1.955929160118103, + "learning_rate": 0.00010561025522220935, + "loss": 0.2653, + "step": 36670 + }, + { + "epoch": 1.4162708984902892, + "grad_norm": 2.2465381622314453, + "learning_rate": 0.00010558451420260757, + "loss": 0.3119, + "step": 36680 + }, + { + "epoch": 1.416657013784316, + "grad_norm": 2.108614921569824, + "learning_rate": 0.00010555877318300577, + "loss": 0.2872, + "step": 36690 + }, + { + "epoch": 1.4170431290783427, + "grad_norm": 1.35493004322052, + "learning_rate": 0.000105533032163404, + "loss": 0.1491, + "step": 36700 + }, + { + "epoch": 1.4174292443723697, + "grad_norm": 0.9102393984794617, + "learning_rate": 0.0001055072911438022, + "loss": 0.1565, + "step": 36710 + }, + { + "epoch": 1.4178153596663963, + "grad_norm": 1.7878345251083374, + "learning_rate": 0.00010548155012420043, + "loss": 0.4864, + "step": 36720 + }, + { + "epoch": 1.4182014749604233, + "grad_norm": 1.4333637952804565, + "learning_rate": 0.00010545580910459864, + "loss": 0.3246, + "step": 36730 + }, + { + "epoch": 1.41858759025445, + "grad_norm": 0.0966360941529274, + "learning_rate": 0.00010543006808499684, + "loss": 0.1758, + "step": 36740 + }, + { + "epoch": 1.4189737055484768, + "grad_norm": 0.30043545365333557, + "learning_rate": 0.00010540432706539507, + "loss": 0.1811, + "step": 36750 + }, + { + "epoch": 1.4193598208425036, + "grad_norm": 1.7705951929092407, + "learning_rate": 0.00010537858604579327, + "loss": 0.1824, + "step": 36760 + }, + { + "epoch": 1.4197459361365303, + "grad_norm": 1.2025195360183716, + "learning_rate": 0.0001053528450261915, + "loss": 0.1463, + "step": 36770 + }, + { + "epoch": 1.420132051430557, + "grad_norm": 0.3154304325580597, + "learning_rate": 0.00010532710400658972, + "loss": 0.1817, + "step": 36780 + }, + { + "epoch": 1.4205181667245839, + "grad_norm": 3.392331838607788, + "learning_rate": 0.00010530136298698792, + "loss": 0.2938, + "step": 36790 + }, + { + "epoch": 1.4209042820186109, + "grad_norm": 0.7256132364273071, + "learning_rate": 0.00010527562196738613, + "loss": 0.28, + "step": 36800 + }, + { + "epoch": 1.4212903973126376, + "grad_norm": 1.9007991552352905, + "learning_rate": 0.00010524988094778435, + "loss": 0.2306, + "step": 36810 + }, + { + "epoch": 1.4216765126066644, + "grad_norm": 3.21189546585083, + "learning_rate": 0.00010522413992818256, + "loss": 0.2945, + "step": 36820 + }, + { + "epoch": 1.4220626279006912, + "grad_norm": 0.20476600527763367, + "learning_rate": 0.00010519839890858076, + "loss": 0.3129, + "step": 36830 + }, + { + "epoch": 1.422448743194718, + "grad_norm": 1.4225107431411743, + "learning_rate": 0.00010517265788897899, + "loss": 0.2262, + "step": 36840 + }, + { + "epoch": 1.4228348584887447, + "grad_norm": 1.203728437423706, + "learning_rate": 0.00010514691686937721, + "loss": 0.176, + "step": 36850 + }, + { + "epoch": 1.4232209737827715, + "grad_norm": 0.3001759648323059, + "learning_rate": 0.00010512117584977541, + "loss": 0.218, + "step": 36860 + }, + { + "epoch": 1.4236070890767984, + "grad_norm": 2.3143389225006104, + "learning_rate": 0.00010509543483017363, + "loss": 0.1562, + "step": 36870 + }, + { + "epoch": 1.423993204370825, + "grad_norm": 0.5087364912033081, + "learning_rate": 0.00010506969381057184, + "loss": 0.1207, + "step": 36880 + }, + { + "epoch": 1.424379319664852, + "grad_norm": 1.6521960496902466, + "learning_rate": 0.00010504395279097005, + "loss": 0.4756, + "step": 36890 + }, + { + "epoch": 1.4247654349588788, + "grad_norm": 0.09236706793308258, + "learning_rate": 0.00010501821177136825, + "loss": 0.1755, + "step": 36900 + }, + { + "epoch": 1.4251515502529055, + "grad_norm": 0.2143094390630722, + "learning_rate": 0.00010499247075176648, + "loss": 0.2126, + "step": 36910 + }, + { + "epoch": 1.4255376655469323, + "grad_norm": 1.210170865058899, + "learning_rate": 0.00010496672973216471, + "loss": 0.1391, + "step": 36920 + }, + { + "epoch": 1.425923780840959, + "grad_norm": 0.6248244047164917, + "learning_rate": 0.00010494098871256291, + "loss": 0.2898, + "step": 36930 + }, + { + "epoch": 1.426309896134986, + "grad_norm": 2.1100337505340576, + "learning_rate": 0.00010491524769296113, + "loss": 0.1198, + "step": 36940 + }, + { + "epoch": 1.4266960114290126, + "grad_norm": 2.5673348903656006, + "learning_rate": 0.00010488950667335933, + "loss": 0.2284, + "step": 36950 + }, + { + "epoch": 1.4270821267230396, + "grad_norm": 1.868195652961731, + "learning_rate": 0.00010486376565375755, + "loss": 0.3738, + "step": 36960 + }, + { + "epoch": 1.4274682420170663, + "grad_norm": 1.0951671600341797, + "learning_rate": 0.00010483802463415577, + "loss": 0.1424, + "step": 36970 + }, + { + "epoch": 1.4278543573110931, + "grad_norm": 0.4791143536567688, + "learning_rate": 0.00010481228361455397, + "loss": 0.1445, + "step": 36980 + }, + { + "epoch": 1.4282404726051199, + "grad_norm": 3.2477540969848633, + "learning_rate": 0.0001047865425949522, + "loss": 0.2676, + "step": 36990 + }, + { + "epoch": 1.4286265878991466, + "grad_norm": 0.8082342147827148, + "learning_rate": 0.0001047608015753504, + "loss": 0.2295, + "step": 37000 + }, + { + "epoch": 1.4290127031931734, + "grad_norm": 2.3048954010009766, + "learning_rate": 0.00010473506055574863, + "loss": 0.1996, + "step": 37010 + }, + { + "epoch": 1.4293988184872002, + "grad_norm": 0.42648783326148987, + "learning_rate": 0.00010470931953614683, + "loss": 0.2549, + "step": 37020 + }, + { + "epoch": 1.4297849337812272, + "grad_norm": 1.8794362545013428, + "learning_rate": 0.00010468357851654504, + "loss": 0.1929, + "step": 37030 + }, + { + "epoch": 1.430171049075254, + "grad_norm": 0.9679039120674133, + "learning_rate": 0.00010465783749694327, + "loss": 0.1544, + "step": 37040 + }, + { + "epoch": 1.4305571643692807, + "grad_norm": 0.7789367437362671, + "learning_rate": 0.00010463209647734147, + "loss": 0.2251, + "step": 37050 + }, + { + "epoch": 1.4309432796633075, + "grad_norm": 1.1705437898635864, + "learning_rate": 0.0001046063554577397, + "loss": 0.1567, + "step": 37060 + }, + { + "epoch": 1.4313293949573342, + "grad_norm": 0.13087430596351624, + "learning_rate": 0.00010458061443813789, + "loss": 0.3237, + "step": 37070 + }, + { + "epoch": 1.431715510251361, + "grad_norm": 1.0658761262893677, + "learning_rate": 0.00010455487341853612, + "loss": 0.24, + "step": 37080 + }, + { + "epoch": 1.4321016255453878, + "grad_norm": 1.686922311782837, + "learning_rate": 0.00010452913239893433, + "loss": 0.1911, + "step": 37090 + }, + { + "epoch": 1.4324877408394148, + "grad_norm": 2.010221481323242, + "learning_rate": 0.00010450339137933253, + "loss": 0.4569, + "step": 37100 + }, + { + "epoch": 1.4328738561334413, + "grad_norm": 0.8007562160491943, + "learning_rate": 0.00010447765035973076, + "loss": 0.1198, + "step": 37110 + }, + { + "epoch": 1.4332599714274683, + "grad_norm": 0.5455211997032166, + "learning_rate": 0.00010445190934012896, + "loss": 0.2179, + "step": 37120 + }, + { + "epoch": 1.433646086721495, + "grad_norm": 1.4133542776107788, + "learning_rate": 0.00010442616832052719, + "loss": 0.45, + "step": 37130 + }, + { + "epoch": 1.4340322020155218, + "grad_norm": 1.6218222379684448, + "learning_rate": 0.00010440042730092539, + "loss": 0.19, + "step": 37140 + }, + { + "epoch": 1.4344183173095486, + "grad_norm": 0.6425970196723938, + "learning_rate": 0.00010437468628132361, + "loss": 0.2345, + "step": 37150 + }, + { + "epoch": 1.4348044326035754, + "grad_norm": 0.4344918131828308, + "learning_rate": 0.00010434894526172183, + "loss": 0.2532, + "step": 37160 + }, + { + "epoch": 1.4351905478976021, + "grad_norm": 0.6628998517990112, + "learning_rate": 0.00010432320424212003, + "loss": 0.2281, + "step": 37170 + }, + { + "epoch": 1.435576663191629, + "grad_norm": 0.8695842027664185, + "learning_rate": 0.00010429746322251825, + "loss": 0.2796, + "step": 37180 + }, + { + "epoch": 1.435962778485656, + "grad_norm": 0.16481854021549225, + "learning_rate": 0.00010427172220291645, + "loss": 0.2333, + "step": 37190 + }, + { + "epoch": 1.4363488937796827, + "grad_norm": 0.7194454073905945, + "learning_rate": 0.00010424598118331468, + "loss": 0.1413, + "step": 37200 + }, + { + "epoch": 1.4367350090737094, + "grad_norm": 4.845508575439453, + "learning_rate": 0.00010422024016371288, + "loss": 0.3944, + "step": 37210 + }, + { + "epoch": 1.4371211243677362, + "grad_norm": 2.6386618614196777, + "learning_rate": 0.0001041944991441111, + "loss": 0.3037, + "step": 37220 + }, + { + "epoch": 1.437507239661763, + "grad_norm": 0.4089922308921814, + "learning_rate": 0.00010416875812450932, + "loss": 0.2985, + "step": 37230 + }, + { + "epoch": 1.4378933549557897, + "grad_norm": 1.456944465637207, + "learning_rate": 0.00010414301710490752, + "loss": 0.3031, + "step": 37240 + }, + { + "epoch": 1.4382794702498165, + "grad_norm": 1.301829218864441, + "learning_rate": 0.00010411727608530575, + "loss": 0.2578, + "step": 37250 + }, + { + "epoch": 1.4386655855438435, + "grad_norm": 1.2072703838348389, + "learning_rate": 0.00010409153506570395, + "loss": 0.391, + "step": 37260 + }, + { + "epoch": 1.4390517008378703, + "grad_norm": 0.5538531541824341, + "learning_rate": 0.00010406579404610217, + "loss": 0.1867, + "step": 37270 + }, + { + "epoch": 1.439437816131897, + "grad_norm": 1.0898678302764893, + "learning_rate": 0.0001040400530265004, + "loss": 0.2112, + "step": 37280 + }, + { + "epoch": 1.4398239314259238, + "grad_norm": 1.5769239664077759, + "learning_rate": 0.0001040143120068986, + "loss": 0.3121, + "step": 37290 + }, + { + "epoch": 1.4402100467199506, + "grad_norm": 0.3964422345161438, + "learning_rate": 0.00010398857098729681, + "loss": 0.196, + "step": 37300 + }, + { + "epoch": 1.4405961620139773, + "grad_norm": 1.0268182754516602, + "learning_rate": 0.00010396282996769501, + "loss": 0.1785, + "step": 37310 + }, + { + "epoch": 1.440982277308004, + "grad_norm": 1.750826358795166, + "learning_rate": 0.00010393708894809324, + "loss": 0.3654, + "step": 37320 + }, + { + "epoch": 1.441368392602031, + "grad_norm": 1.1231745481491089, + "learning_rate": 0.00010391134792849144, + "loss": 0.2594, + "step": 37330 + }, + { + "epoch": 1.4417545078960576, + "grad_norm": 0.2897786498069763, + "learning_rate": 0.00010388560690888967, + "loss": 0.3483, + "step": 37340 + }, + { + "epoch": 1.4421406231900846, + "grad_norm": 0.07170752435922623, + "learning_rate": 0.00010385986588928789, + "loss": 0.1462, + "step": 37350 + }, + { + "epoch": 1.4425267384841114, + "grad_norm": 3.326099395751953, + "learning_rate": 0.00010383412486968609, + "loss": 0.3785, + "step": 37360 + }, + { + "epoch": 1.4429128537781382, + "grad_norm": 0.5063263773918152, + "learning_rate": 0.0001038083838500843, + "loss": 0.2974, + "step": 37370 + }, + { + "epoch": 1.443298969072165, + "grad_norm": 1.160088062286377, + "learning_rate": 0.00010378264283048252, + "loss": 0.3003, + "step": 37380 + }, + { + "epoch": 1.4436850843661917, + "grad_norm": 0.7043284177780151, + "learning_rate": 0.00010375690181088073, + "loss": 0.3193, + "step": 37390 + }, + { + "epoch": 1.4440711996602185, + "grad_norm": 0.2916620373725891, + "learning_rate": 0.00010373116079127896, + "loss": 0.235, + "step": 37400 + }, + { + "epoch": 1.4444573149542452, + "grad_norm": 2.1940219402313232, + "learning_rate": 0.00010370541977167716, + "loss": 0.2657, + "step": 37410 + }, + { + "epoch": 1.4448434302482722, + "grad_norm": 1.849794626235962, + "learning_rate": 0.00010367967875207539, + "loss": 0.3121, + "step": 37420 + }, + { + "epoch": 1.445229545542299, + "grad_norm": 1.1139589548110962, + "learning_rate": 0.00010365393773247359, + "loss": 0.2179, + "step": 37430 + }, + { + "epoch": 1.4456156608363258, + "grad_norm": 0.917142927646637, + "learning_rate": 0.0001036281967128718, + "loss": 0.2406, + "step": 37440 + }, + { + "epoch": 1.4460017761303525, + "grad_norm": 1.6973673105239868, + "learning_rate": 0.00010360245569327001, + "loss": 0.3236, + "step": 37450 + }, + { + "epoch": 1.4463878914243793, + "grad_norm": 0.3979933559894562, + "learning_rate": 0.00010357671467366823, + "loss": 0.1477, + "step": 37460 + }, + { + "epoch": 1.446774006718406, + "grad_norm": 0.8938451409339905, + "learning_rate": 0.00010355097365406645, + "loss": 0.116, + "step": 37470 + }, + { + "epoch": 1.4471601220124328, + "grad_norm": 1.068787693977356, + "learning_rate": 0.00010352523263446465, + "loss": 0.2342, + "step": 37480 + }, + { + "epoch": 1.4475462373064598, + "grad_norm": 1.1985591650009155, + "learning_rate": 0.00010349949161486288, + "loss": 0.2423, + "step": 37490 + }, + { + "epoch": 1.4479323526004866, + "grad_norm": 1.0383973121643066, + "learning_rate": 0.00010347375059526108, + "loss": 0.3059, + "step": 37500 + }, + { + "epoch": 1.4483184678945134, + "grad_norm": 0.26652297377586365, + "learning_rate": 0.0001034480095756593, + "loss": 0.262, + "step": 37510 + }, + { + "epoch": 1.4487045831885401, + "grad_norm": 1.8498083353042603, + "learning_rate": 0.0001034222685560575, + "loss": 0.2122, + "step": 37520 + }, + { + "epoch": 1.4490906984825669, + "grad_norm": 1.2896068096160889, + "learning_rate": 0.00010339652753645572, + "loss": 0.2939, + "step": 37530 + }, + { + "epoch": 1.4494768137765937, + "grad_norm": 1.423343300819397, + "learning_rate": 0.00010337078651685395, + "loss": 0.3375, + "step": 37540 + }, + { + "epoch": 1.4498629290706204, + "grad_norm": 1.8248246908187866, + "learning_rate": 0.00010334504549725215, + "loss": 0.3534, + "step": 37550 + }, + { + "epoch": 1.4502490443646474, + "grad_norm": 2.3713393211364746, + "learning_rate": 0.00010331930447765037, + "loss": 0.2795, + "step": 37560 + }, + { + "epoch": 1.450635159658674, + "grad_norm": 2.1431849002838135, + "learning_rate": 0.00010329356345804857, + "loss": 0.2503, + "step": 37570 + }, + { + "epoch": 1.451021274952701, + "grad_norm": 1.6521297693252563, + "learning_rate": 0.0001032678224384468, + "loss": 0.1789, + "step": 37580 + }, + { + "epoch": 1.4514073902467277, + "grad_norm": 1.5589754581451416, + "learning_rate": 0.00010324208141884501, + "loss": 0.3201, + "step": 37590 + }, + { + "epoch": 1.4517935055407545, + "grad_norm": 0.4339803159236908, + "learning_rate": 0.00010321634039924321, + "loss": 0.2964, + "step": 37600 + }, + { + "epoch": 1.4521796208347812, + "grad_norm": 0.30054792761802673, + "learning_rate": 0.00010319059937964144, + "loss": 0.3477, + "step": 37610 + }, + { + "epoch": 1.452565736128808, + "grad_norm": 2.2864038944244385, + "learning_rate": 0.00010316485836003964, + "loss": 0.2849, + "step": 37620 + }, + { + "epoch": 1.4529518514228348, + "grad_norm": 1.8392651081085205, + "learning_rate": 0.00010313911734043787, + "loss": 0.2089, + "step": 37630 + }, + { + "epoch": 1.4533379667168616, + "grad_norm": 1.0444347858428955, + "learning_rate": 0.00010311337632083607, + "loss": 0.2306, + "step": 37640 + }, + { + "epoch": 1.4537240820108885, + "grad_norm": 1.265647053718567, + "learning_rate": 0.00010308763530123429, + "loss": 0.2192, + "step": 37650 + }, + { + "epoch": 1.4541101973049153, + "grad_norm": 1.0464913845062256, + "learning_rate": 0.0001030618942816325, + "loss": 0.2857, + "step": 37660 + }, + { + "epoch": 1.454496312598942, + "grad_norm": 0.5931240320205688, + "learning_rate": 0.0001030361532620307, + "loss": 0.1751, + "step": 37670 + }, + { + "epoch": 1.4548824278929688, + "grad_norm": 2.0155346393585205, + "learning_rate": 0.00010301041224242893, + "loss": 0.4124, + "step": 37680 + }, + { + "epoch": 1.4552685431869956, + "grad_norm": 2.843345880508423, + "learning_rate": 0.00010298467122282713, + "loss": 0.2119, + "step": 37690 + }, + { + "epoch": 1.4556546584810224, + "grad_norm": 1.977612018585205, + "learning_rate": 0.00010295893020322536, + "loss": 0.271, + "step": 37700 + }, + { + "epoch": 1.4560407737750491, + "grad_norm": 1.2862237691879272, + "learning_rate": 0.00010293318918362356, + "loss": 0.4147, + "step": 37710 + }, + { + "epoch": 1.4564268890690761, + "grad_norm": 0.8235340118408203, + "learning_rate": 0.00010290744816402179, + "loss": 0.1918, + "step": 37720 + }, + { + "epoch": 1.456813004363103, + "grad_norm": 1.0351753234863281, + "learning_rate": 0.00010288170714442, + "loss": 0.2343, + "step": 37730 + }, + { + "epoch": 1.4571991196571297, + "grad_norm": 0.2567160725593567, + "learning_rate": 0.0001028559661248182, + "loss": 0.2664, + "step": 37740 + }, + { + "epoch": 1.4575852349511564, + "grad_norm": 2.7498669624328613, + "learning_rate": 0.00010283022510521643, + "loss": 0.2839, + "step": 37750 + }, + { + "epoch": 1.4579713502451832, + "grad_norm": 0.7125422358512878, + "learning_rate": 0.00010280448408561463, + "loss": 0.2411, + "step": 37760 + }, + { + "epoch": 1.45835746553921, + "grad_norm": 0.720761239528656, + "learning_rate": 0.00010277874306601285, + "loss": 0.315, + "step": 37770 + }, + { + "epoch": 1.4587435808332367, + "grad_norm": 3.188563346862793, + "learning_rate": 0.00010275300204641108, + "loss": 0.3945, + "step": 37780 + }, + { + "epoch": 1.4591296961272637, + "grad_norm": 1.23422110080719, + "learning_rate": 0.00010272726102680928, + "loss": 0.1828, + "step": 37790 + }, + { + "epoch": 1.4595158114212903, + "grad_norm": 1.1572456359863281, + "learning_rate": 0.00010270152000720749, + "loss": 0.1687, + "step": 37800 + }, + { + "epoch": 1.4599019267153173, + "grad_norm": 0.6565262079238892, + "learning_rate": 0.00010267577898760569, + "loss": 0.3145, + "step": 37810 + }, + { + "epoch": 1.460288042009344, + "grad_norm": 0.6239646673202515, + "learning_rate": 0.00010265003796800392, + "loss": 0.2381, + "step": 37820 + }, + { + "epoch": 1.4606741573033708, + "grad_norm": 0.7578912377357483, + "learning_rate": 0.00010262429694840212, + "loss": 0.2019, + "step": 37830 + }, + { + "epoch": 1.4610602725973976, + "grad_norm": 1.0978549718856812, + "learning_rate": 0.00010259855592880035, + "loss": 0.2477, + "step": 37840 + }, + { + "epoch": 1.4614463878914243, + "grad_norm": 0.3615519404411316, + "learning_rate": 0.00010257281490919857, + "loss": 0.4241, + "step": 37850 + }, + { + "epoch": 1.461832503185451, + "grad_norm": 0.10355047881603241, + "learning_rate": 0.00010254707388959677, + "loss": 0.1181, + "step": 37860 + }, + { + "epoch": 1.4622186184794779, + "grad_norm": 0.23957425355911255, + "learning_rate": 0.00010252133286999499, + "loss": 0.1675, + "step": 37870 + }, + { + "epoch": 1.4626047337735049, + "grad_norm": 1.7581062316894531, + "learning_rate": 0.00010249559185039318, + "loss": 0.241, + "step": 37880 + }, + { + "epoch": 1.4629908490675316, + "grad_norm": 1.660989761352539, + "learning_rate": 0.00010246985083079141, + "loss": 0.2099, + "step": 37890 + }, + { + "epoch": 1.4633769643615584, + "grad_norm": 1.1359142065048218, + "learning_rate": 0.00010244410981118964, + "loss": 0.3686, + "step": 37900 + }, + { + "epoch": 1.4637630796555852, + "grad_norm": 0.9489149451255798, + "learning_rate": 0.00010241836879158784, + "loss": 0.3616, + "step": 37910 + }, + { + "epoch": 1.464149194949612, + "grad_norm": 1.5209956169128418, + "learning_rate": 0.00010239262777198607, + "loss": 0.3159, + "step": 37920 + }, + { + "epoch": 1.4645353102436387, + "grad_norm": 0.848943293094635, + "learning_rate": 0.00010236688675238427, + "loss": 0.3228, + "step": 37930 + }, + { + "epoch": 1.4649214255376655, + "grad_norm": 1.1476777791976929, + "learning_rate": 0.00010234114573278248, + "loss": 0.1873, + "step": 37940 + }, + { + "epoch": 1.4653075408316925, + "grad_norm": 1.4093862771987915, + "learning_rate": 0.00010231540471318068, + "loss": 0.5554, + "step": 37950 + }, + { + "epoch": 1.465693656125719, + "grad_norm": 0.17349161207675934, + "learning_rate": 0.0001022896636935789, + "loss": 0.2992, + "step": 37960 + }, + { + "epoch": 1.466079771419746, + "grad_norm": 0.28885993361473083, + "learning_rate": 0.00010226392267397713, + "loss": 0.2618, + "step": 37970 + }, + { + "epoch": 1.4664658867137728, + "grad_norm": 1.1087830066680908, + "learning_rate": 0.00010223818165437533, + "loss": 0.2647, + "step": 37980 + }, + { + "epoch": 1.4668520020077995, + "grad_norm": 0.08338876068592072, + "learning_rate": 0.00010221244063477356, + "loss": 0.3057, + "step": 37990 + }, + { + "epoch": 1.4672381173018263, + "grad_norm": 2.159362316131592, + "learning_rate": 0.00010218669961517176, + "loss": 0.1499, + "step": 38000 + }, + { + "epoch": 1.467624232595853, + "grad_norm": 0.8207988142967224, + "learning_rate": 0.00010216095859556997, + "loss": 0.1482, + "step": 38010 + }, + { + "epoch": 1.46801034788988, + "grad_norm": 0.5458611845970154, + "learning_rate": 0.00010213521757596818, + "loss": 0.2794, + "step": 38020 + }, + { + "epoch": 1.4683964631839066, + "grad_norm": 1.6955047845840454, + "learning_rate": 0.0001021094765563664, + "loss": 0.2627, + "step": 38030 + }, + { + "epoch": 1.4687825784779336, + "grad_norm": 0.9796440601348877, + "learning_rate": 0.00010208373553676463, + "loss": 0.1924, + "step": 38040 + }, + { + "epoch": 1.4691686937719604, + "grad_norm": 0.9906508326530457, + "learning_rate": 0.00010205799451716282, + "loss": 0.1597, + "step": 38050 + }, + { + "epoch": 1.4695548090659871, + "grad_norm": 1.8590656518936157, + "learning_rate": 0.00010203225349756105, + "loss": 0.2874, + "step": 38060 + }, + { + "epoch": 1.469940924360014, + "grad_norm": 1.6403672695159912, + "learning_rate": 0.00010200651247795925, + "loss": 0.2926, + "step": 38070 + }, + { + "epoch": 1.4703270396540407, + "grad_norm": 0.4410895109176636, + "learning_rate": 0.00010198077145835746, + "loss": 0.1368, + "step": 38080 + }, + { + "epoch": 1.4707131549480674, + "grad_norm": 2.4955286979675293, + "learning_rate": 0.00010195503043875569, + "loss": 0.3195, + "step": 38090 + }, + { + "epoch": 1.4710992702420942, + "grad_norm": 1.1799029111862183, + "learning_rate": 0.00010192928941915389, + "loss": 0.1867, + "step": 38100 + }, + { + "epoch": 1.4714853855361212, + "grad_norm": 0.7959389090538025, + "learning_rate": 0.00010190354839955212, + "loss": 0.2988, + "step": 38110 + }, + { + "epoch": 1.471871500830148, + "grad_norm": 2.7750720977783203, + "learning_rate": 0.00010187780737995032, + "loss": 0.1886, + "step": 38120 + }, + { + "epoch": 1.4722576161241747, + "grad_norm": 1.5834373235702515, + "learning_rate": 0.00010185206636034854, + "loss": 0.3542, + "step": 38130 + }, + { + "epoch": 1.4726437314182015, + "grad_norm": 1.9757747650146484, + "learning_rate": 0.00010182632534074674, + "loss": 0.302, + "step": 38140 + }, + { + "epoch": 1.4730298467122283, + "grad_norm": 1.1752204895019531, + "learning_rate": 0.00010180058432114497, + "loss": 0.3535, + "step": 38150 + }, + { + "epoch": 1.473415962006255, + "grad_norm": 0.3877789378166199, + "learning_rate": 0.00010177484330154318, + "loss": 0.2508, + "step": 38160 + }, + { + "epoch": 1.4738020773002818, + "grad_norm": 0.13749545812606812, + "learning_rate": 0.00010174910228194138, + "loss": 0.2141, + "step": 38170 + }, + { + "epoch": 1.4741881925943088, + "grad_norm": 1.3663641214370728, + "learning_rate": 0.00010172336126233961, + "loss": 0.3231, + "step": 38180 + }, + { + "epoch": 1.4745743078883353, + "grad_norm": 1.6267393827438354, + "learning_rate": 0.00010169762024273781, + "loss": 0.3233, + "step": 38190 + }, + { + "epoch": 1.4749604231823623, + "grad_norm": 0.2993789315223694, + "learning_rate": 0.00010167187922313604, + "loss": 0.28, + "step": 38200 + }, + { + "epoch": 1.475346538476389, + "grad_norm": 0.16693222522735596, + "learning_rate": 0.00010164613820353424, + "loss": 0.188, + "step": 38210 + }, + { + "epoch": 1.4757326537704158, + "grad_norm": 0.6939979791641235, + "learning_rate": 0.00010162039718393246, + "loss": 0.263, + "step": 38220 + }, + { + "epoch": 1.4761187690644426, + "grad_norm": 0.37910985946655273, + "learning_rate": 0.00010159465616433068, + "loss": 0.1963, + "step": 38230 + }, + { + "epoch": 1.4765048843584694, + "grad_norm": 1.782188892364502, + "learning_rate": 0.00010156891514472888, + "loss": 0.3814, + "step": 38240 + }, + { + "epoch": 1.4768909996524964, + "grad_norm": 1.159278392791748, + "learning_rate": 0.0001015431741251271, + "loss": 0.2043, + "step": 38250 + }, + { + "epoch": 1.477277114946523, + "grad_norm": 1.09486985206604, + "learning_rate": 0.0001015174331055253, + "loss": 0.2128, + "step": 38260 + }, + { + "epoch": 1.47766323024055, + "grad_norm": 0.36655205488204956, + "learning_rate": 0.00010149169208592353, + "loss": 0.298, + "step": 38270 + }, + { + "epoch": 1.4780493455345767, + "grad_norm": 0.8908851742744446, + "learning_rate": 0.00010146595106632176, + "loss": 0.3707, + "step": 38280 + }, + { + "epoch": 1.4784354608286034, + "grad_norm": 0.251338928937912, + "learning_rate": 0.00010144021004671996, + "loss": 0.2495, + "step": 38290 + }, + { + "epoch": 1.4788215761226302, + "grad_norm": 1.0613712072372437, + "learning_rate": 0.00010141446902711817, + "loss": 0.2112, + "step": 38300 + }, + { + "epoch": 1.479207691416657, + "grad_norm": 1.459799885749817, + "learning_rate": 0.00010138872800751637, + "loss": 0.2595, + "step": 38310 + }, + { + "epoch": 1.4795938067106837, + "grad_norm": 2.6898603439331055, + "learning_rate": 0.0001013629869879146, + "loss": 0.2758, + "step": 38320 + }, + { + "epoch": 1.4799799220047105, + "grad_norm": 0.19628773629665375, + "learning_rate": 0.0001013372459683128, + "loss": 0.1843, + "step": 38330 + }, + { + "epoch": 1.4803660372987375, + "grad_norm": 2.0871078968048096, + "learning_rate": 0.00010131150494871102, + "loss": 0.1661, + "step": 38340 + }, + { + "epoch": 1.4807521525927643, + "grad_norm": 0.7689336538314819, + "learning_rate": 0.00010128576392910925, + "loss": 0.157, + "step": 38350 + }, + { + "epoch": 1.481138267886791, + "grad_norm": 1.4471644163131714, + "learning_rate": 0.00010126002290950745, + "loss": 0.2159, + "step": 38360 + }, + { + "epoch": 1.4815243831808178, + "grad_norm": 2.198559522628784, + "learning_rate": 0.00010123428188990566, + "loss": 0.5609, + "step": 38370 + }, + { + "epoch": 1.4819104984748446, + "grad_norm": 0.16012130677700043, + "learning_rate": 0.00010120854087030386, + "loss": 0.1979, + "step": 38380 + }, + { + "epoch": 1.4822966137688713, + "grad_norm": 1.0222225189208984, + "learning_rate": 0.00010118279985070209, + "loss": 0.1547, + "step": 38390 + }, + { + "epoch": 1.482682729062898, + "grad_norm": 2.7192416191101074, + "learning_rate": 0.00010115705883110032, + "loss": 0.4942, + "step": 38400 + }, + { + "epoch": 1.483068844356925, + "grad_norm": 1.887128472328186, + "learning_rate": 0.00010113131781149852, + "loss": 0.177, + "step": 38410 + }, + { + "epoch": 1.4834549596509516, + "grad_norm": 2.7628560066223145, + "learning_rate": 0.00010110557679189674, + "loss": 0.2931, + "step": 38420 + }, + { + "epoch": 1.4838410749449786, + "grad_norm": 0.3852572739124298, + "learning_rate": 0.00010107983577229494, + "loss": 0.3392, + "step": 38430 + }, + { + "epoch": 1.4842271902390054, + "grad_norm": 1.047448992729187, + "learning_rate": 0.00010105409475269316, + "loss": 0.3741, + "step": 38440 + }, + { + "epoch": 1.4846133055330322, + "grad_norm": 1.4930602312088013, + "learning_rate": 0.00010102835373309136, + "loss": 0.2564, + "step": 38450 + }, + { + "epoch": 1.484999420827059, + "grad_norm": 1.3012608289718628, + "learning_rate": 0.00010100261271348958, + "loss": 0.3376, + "step": 38460 + }, + { + "epoch": 1.4853855361210857, + "grad_norm": 2.163942337036133, + "learning_rate": 0.00010097687169388781, + "loss": 0.3548, + "step": 38470 + }, + { + "epoch": 1.4857716514151125, + "grad_norm": 1.864189624786377, + "learning_rate": 0.00010095113067428601, + "loss": 0.165, + "step": 38480 + }, + { + "epoch": 1.4861577667091392, + "grad_norm": 0.5661312341690063, + "learning_rate": 0.00010092538965468424, + "loss": 0.1764, + "step": 38490 + }, + { + "epoch": 1.4865438820031662, + "grad_norm": 0.13517481088638306, + "learning_rate": 0.00010089964863508244, + "loss": 0.5223, + "step": 38500 + }, + { + "epoch": 1.486929997297193, + "grad_norm": 0.665143609046936, + "learning_rate": 0.00010087390761548065, + "loss": 0.1943, + "step": 38510 + }, + { + "epoch": 1.4873161125912198, + "grad_norm": 1.2759610414505005, + "learning_rate": 0.00010084816659587885, + "loss": 0.3023, + "step": 38520 + }, + { + "epoch": 1.4877022278852465, + "grad_norm": 1.3209573030471802, + "learning_rate": 0.00010082242557627708, + "loss": 0.1101, + "step": 38530 + }, + { + "epoch": 1.4880883431792733, + "grad_norm": 1.2501552104949951, + "learning_rate": 0.0001007966845566753, + "loss": 0.3931, + "step": 38540 + }, + { + "epoch": 1.4884744584733, + "grad_norm": 0.6862074732780457, + "learning_rate": 0.0001007709435370735, + "loss": 0.4093, + "step": 38550 + }, + { + "epoch": 1.4888605737673268, + "grad_norm": 1.90501070022583, + "learning_rate": 0.00010074520251747173, + "loss": 0.2553, + "step": 38560 + }, + { + "epoch": 1.4892466890613538, + "grad_norm": 1.6547000408172607, + "learning_rate": 0.00010071946149786993, + "loss": 0.1558, + "step": 38570 + }, + { + "epoch": 1.4896328043553806, + "grad_norm": 0.35097751021385193, + "learning_rate": 0.00010069372047826814, + "loss": 0.2253, + "step": 38580 + }, + { + "epoch": 1.4900189196494074, + "grad_norm": 0.15141837298870087, + "learning_rate": 0.00010066797945866637, + "loss": 0.2124, + "step": 38590 + }, + { + "epoch": 1.4904050349434341, + "grad_norm": 1.7070786952972412, + "learning_rate": 0.00010064223843906457, + "loss": 0.3181, + "step": 38600 + }, + { + "epoch": 1.490791150237461, + "grad_norm": 0.5400305390357971, + "learning_rate": 0.0001006164974194628, + "loss": 0.3203, + "step": 38610 + }, + { + "epoch": 1.4911772655314877, + "grad_norm": 1.6475050449371338, + "learning_rate": 0.000100590756399861, + "loss": 0.2965, + "step": 38620 + }, + { + "epoch": 1.4915633808255144, + "grad_norm": 0.21372176706790924, + "learning_rate": 0.00010056501538025922, + "loss": 0.1456, + "step": 38630 + }, + { + "epoch": 1.4919494961195414, + "grad_norm": 0.24179309606552124, + "learning_rate": 0.00010053927436065742, + "loss": 0.2077, + "step": 38640 + }, + { + "epoch": 1.492335611413568, + "grad_norm": 1.2079945802688599, + "learning_rate": 0.00010051353334105564, + "loss": 0.1554, + "step": 38650 + }, + { + "epoch": 1.492721726707595, + "grad_norm": 1.8915836811065674, + "learning_rate": 0.00010048779232145386, + "loss": 0.3815, + "step": 38660 + }, + { + "epoch": 1.4931078420016217, + "grad_norm": 1.8128750324249268, + "learning_rate": 0.00010046205130185206, + "loss": 0.1944, + "step": 38670 + }, + { + "epoch": 1.4934939572956485, + "grad_norm": 2.0955018997192383, + "learning_rate": 0.00010043631028225029, + "loss": 0.2468, + "step": 38680 + }, + { + "epoch": 1.4938800725896753, + "grad_norm": 3.0578064918518066, + "learning_rate": 0.00010041056926264849, + "loss": 0.2186, + "step": 38690 + }, + { + "epoch": 1.494266187883702, + "grad_norm": 0.7817699909210205, + "learning_rate": 0.00010038482824304672, + "loss": 0.1154, + "step": 38700 + }, + { + "epoch": 1.4946523031777288, + "grad_norm": 6.556485176086426, + "learning_rate": 0.00010035908722344493, + "loss": 0.4752, + "step": 38710 + }, + { + "epoch": 1.4950384184717556, + "grad_norm": 2.1970055103302, + "learning_rate": 0.00010033334620384314, + "loss": 0.3234, + "step": 38720 + }, + { + "epoch": 1.4954245337657825, + "grad_norm": 1.8929531574249268, + "learning_rate": 0.00010030760518424136, + "loss": 0.1605, + "step": 38730 + }, + { + "epoch": 1.4958106490598093, + "grad_norm": 0.3992670178413391, + "learning_rate": 0.00010028186416463956, + "loss": 0.2417, + "step": 38740 + }, + { + "epoch": 1.496196764353836, + "grad_norm": 1.7508872747421265, + "learning_rate": 0.00010025612314503778, + "loss": 0.329, + "step": 38750 + }, + { + "epoch": 1.4965828796478629, + "grad_norm": 0.24950659275054932, + "learning_rate": 0.00010023038212543598, + "loss": 0.1159, + "step": 38760 + }, + { + "epoch": 1.4969689949418896, + "grad_norm": 1.5187748670578003, + "learning_rate": 0.00010020464110583421, + "loss": 0.1827, + "step": 38770 + }, + { + "epoch": 1.4973551102359164, + "grad_norm": 1.1223959922790527, + "learning_rate": 0.00010017890008623242, + "loss": 0.1935, + "step": 38780 + }, + { + "epoch": 1.4977412255299432, + "grad_norm": 0.9355156421661377, + "learning_rate": 0.00010015315906663064, + "loss": 0.2225, + "step": 38790 + }, + { + "epoch": 1.4981273408239701, + "grad_norm": 0.4834296405315399, + "learning_rate": 0.00010012741804702885, + "loss": 0.3077, + "step": 38800 + }, + { + "epoch": 1.498513456117997, + "grad_norm": 1.0983386039733887, + "learning_rate": 0.00010010167702742705, + "loss": 0.183, + "step": 38810 + }, + { + "epoch": 1.4988995714120237, + "grad_norm": 0.8350847959518433, + "learning_rate": 0.00010007593600782528, + "loss": 0.3972, + "step": 38820 + }, + { + "epoch": 1.4992856867060504, + "grad_norm": 0.8200152516365051, + "learning_rate": 0.00010005019498822348, + "loss": 0.2043, + "step": 38830 + }, + { + "epoch": 1.4996718020000772, + "grad_norm": 0.9136185050010681, + "learning_rate": 0.0001000244539686217, + "loss": 0.1727, + "step": 38840 + }, + { + "epoch": 1.500057917294104, + "grad_norm": 0.8466988205909729, + "learning_rate": 9.999871294901992e-05, + "loss": 0.2119, + "step": 38850 + }, + { + "epoch": 1.5004440325881307, + "grad_norm": 0.4185144305229187, + "learning_rate": 9.997297192941813e-05, + "loss": 0.4046, + "step": 38860 + }, + { + "epoch": 1.5008301478821577, + "grad_norm": 2.232264518737793, + "learning_rate": 9.994723090981634e-05, + "loss": 0.304, + "step": 38870 + }, + { + "epoch": 1.5012162631761843, + "grad_norm": 0.13785889744758606, + "learning_rate": 9.992148989021456e-05, + "loss": 0.1045, + "step": 38880 + }, + { + "epoch": 1.5016023784702113, + "grad_norm": 1.6270711421966553, + "learning_rate": 9.989574887061277e-05, + "loss": 0.2791, + "step": 38890 + }, + { + "epoch": 1.501988493764238, + "grad_norm": 0.08486157655715942, + "learning_rate": 9.987000785101098e-05, + "loss": 0.178, + "step": 38900 + }, + { + "epoch": 1.5023746090582648, + "grad_norm": 1.862197995185852, + "learning_rate": 9.98442668314092e-05, + "loss": 0.2461, + "step": 38910 + }, + { + "epoch": 1.5027607243522916, + "grad_norm": 2.752070903778076, + "learning_rate": 9.981852581180741e-05, + "loss": 0.4299, + "step": 38920 + }, + { + "epoch": 1.5031468396463183, + "grad_norm": 2.08542537689209, + "learning_rate": 9.979278479220562e-05, + "loss": 0.2898, + "step": 38930 + }, + { + "epoch": 1.5035329549403453, + "grad_norm": 0.8629382848739624, + "learning_rate": 9.976704377260384e-05, + "loss": 0.226, + "step": 38940 + }, + { + "epoch": 1.5039190702343719, + "grad_norm": 0.5178211331367493, + "learning_rate": 9.974130275300205e-05, + "loss": 0.2444, + "step": 38950 + }, + { + "epoch": 1.5043051855283989, + "grad_norm": 0.25908491015434265, + "learning_rate": 9.971556173340026e-05, + "loss": 0.1643, + "step": 38960 + }, + { + "epoch": 1.5046913008224254, + "grad_norm": 1.1818209886550903, + "learning_rate": 9.968982071379848e-05, + "loss": 0.3187, + "step": 38970 + }, + { + "epoch": 1.5050774161164524, + "grad_norm": 0.13186976313591003, + "learning_rate": 9.966407969419669e-05, + "loss": 0.1982, + "step": 38980 + }, + { + "epoch": 1.5054635314104792, + "grad_norm": 0.18049825727939606, + "learning_rate": 9.963833867459492e-05, + "loss": 0.1288, + "step": 38990 + }, + { + "epoch": 1.505849646704506, + "grad_norm": 0.30261853337287903, + "learning_rate": 9.961259765499312e-05, + "loss": 0.1704, + "step": 39000 + }, + { + "epoch": 1.5062357619985327, + "grad_norm": 2.1437973976135254, + "learning_rate": 9.958685663539133e-05, + "loss": 0.1272, + "step": 39010 + }, + { + "epoch": 1.5066218772925595, + "grad_norm": 2.2844271659851074, + "learning_rate": 9.956111561578954e-05, + "loss": 0.1314, + "step": 39020 + }, + { + "epoch": 1.5070079925865865, + "grad_norm": 1.5845297574996948, + "learning_rate": 9.953537459618776e-05, + "loss": 0.2023, + "step": 39030 + }, + { + "epoch": 1.507394107880613, + "grad_norm": 0.7256748676300049, + "learning_rate": 9.950963357658597e-05, + "loss": 0.4165, + "step": 39040 + }, + { + "epoch": 1.50778022317464, + "grad_norm": 1.7597005367279053, + "learning_rate": 9.948389255698418e-05, + "loss": 0.4209, + "step": 39050 + }, + { + "epoch": 1.5081663384686668, + "grad_norm": 1.5487171411514282, + "learning_rate": 9.945815153738241e-05, + "loss": 0.1918, + "step": 39060 + }, + { + "epoch": 1.5085524537626935, + "grad_norm": 0.1656871736049652, + "learning_rate": 9.943241051778062e-05, + "loss": 0.3513, + "step": 39070 + }, + { + "epoch": 1.5089385690567203, + "grad_norm": 0.1451992392539978, + "learning_rate": 9.940666949817882e-05, + "loss": 0.1922, + "step": 39080 + }, + { + "epoch": 1.509324684350747, + "grad_norm": 1.1572967767715454, + "learning_rate": 9.938092847857704e-05, + "loss": 0.4263, + "step": 39090 + }, + { + "epoch": 1.509710799644774, + "grad_norm": 0.8721522092819214, + "learning_rate": 9.935518745897525e-05, + "loss": 0.3499, + "step": 39100 + }, + { + "epoch": 1.5100969149388006, + "grad_norm": 2.8486688137054443, + "learning_rate": 9.932944643937346e-05, + "loss": 0.3567, + "step": 39110 + }, + { + "epoch": 1.5104830302328276, + "grad_norm": 1.1864535808563232, + "learning_rate": 9.930370541977169e-05, + "loss": 0.2546, + "step": 39120 + }, + { + "epoch": 1.5108691455268544, + "grad_norm": 0.4636247158050537, + "learning_rate": 9.92779644001699e-05, + "loss": 0.2761, + "step": 39130 + }, + { + "epoch": 1.5112552608208811, + "grad_norm": 0.6326389908790588, + "learning_rate": 9.925222338056812e-05, + "loss": 0.252, + "step": 39140 + }, + { + "epoch": 1.511641376114908, + "grad_norm": 2.1110761165618896, + "learning_rate": 9.922648236096632e-05, + "loss": 0.5427, + "step": 39150 + }, + { + "epoch": 1.5120274914089347, + "grad_norm": 1.0927456617355347, + "learning_rate": 9.920074134136453e-05, + "loss": 0.3336, + "step": 39160 + }, + { + "epoch": 1.5124136067029617, + "grad_norm": 3.6224989891052246, + "learning_rate": 9.917500032176274e-05, + "loss": 0.3689, + "step": 39170 + }, + { + "epoch": 1.5127997219969882, + "grad_norm": 2.422492265701294, + "learning_rate": 9.914925930216097e-05, + "loss": 0.3924, + "step": 39180 + }, + { + "epoch": 1.5131858372910152, + "grad_norm": 0.22901678085327148, + "learning_rate": 9.912351828255918e-05, + "loss": 0.3082, + "step": 39190 + }, + { + "epoch": 1.5135719525850417, + "grad_norm": 0.581598162651062, + "learning_rate": 9.90977772629574e-05, + "loss": 0.3007, + "step": 39200 + }, + { + "epoch": 1.5139580678790687, + "grad_norm": 0.7196664214134216, + "learning_rate": 9.907203624335561e-05, + "loss": 0.2425, + "step": 39210 + }, + { + "epoch": 1.5143441831730955, + "grad_norm": 2.5246760845184326, + "learning_rate": 9.904629522375381e-05, + "loss": 0.3085, + "step": 39220 + }, + { + "epoch": 1.5147302984671223, + "grad_norm": 1.8397210836410522, + "learning_rate": 9.902055420415202e-05, + "loss": 0.2388, + "step": 39230 + }, + { + "epoch": 1.515116413761149, + "grad_norm": 0.5940410494804382, + "learning_rate": 9.899481318455025e-05, + "loss": 0.1184, + "step": 39240 + }, + { + "epoch": 1.5155025290551758, + "grad_norm": 1.5379250049591064, + "learning_rate": 9.896907216494846e-05, + "loss": 0.3253, + "step": 39250 + }, + { + "epoch": 1.5158886443492028, + "grad_norm": 0.8007088303565979, + "learning_rate": 9.894333114534668e-05, + "loss": 0.3057, + "step": 39260 + }, + { + "epoch": 1.5162747596432293, + "grad_norm": 0.9321600198745728, + "learning_rate": 9.891759012574489e-05, + "loss": 0.3249, + "step": 39270 + }, + { + "epoch": 1.5166608749372563, + "grad_norm": 1.5519977807998657, + "learning_rate": 9.88918491061431e-05, + "loss": 0.3368, + "step": 39280 + }, + { + "epoch": 1.517046990231283, + "grad_norm": 0.37695613503456116, + "learning_rate": 9.88661080865413e-05, + "loss": 0.1305, + "step": 39290 + }, + { + "epoch": 1.5174331055253099, + "grad_norm": 1.9956984519958496, + "learning_rate": 9.884036706693952e-05, + "loss": 0.4854, + "step": 39300 + }, + { + "epoch": 1.5178192208193366, + "grad_norm": 1.6110823154449463, + "learning_rate": 9.881462604733774e-05, + "loss": 0.2504, + "step": 39310 + }, + { + "epoch": 1.5182053361133634, + "grad_norm": 0.41702982783317566, + "learning_rate": 9.878888502773596e-05, + "loss": 0.1532, + "step": 39320 + }, + { + "epoch": 1.5185914514073904, + "grad_norm": 2.3595950603485107, + "learning_rate": 9.876314400813417e-05, + "loss": 0.309, + "step": 39330 + }, + { + "epoch": 1.518977566701417, + "grad_norm": 1.1045889854431152, + "learning_rate": 9.873740298853238e-05, + "loss": 0.2858, + "step": 39340 + }, + { + "epoch": 1.519363681995444, + "grad_norm": 1.4641762971878052, + "learning_rate": 9.87116619689306e-05, + "loss": 0.3159, + "step": 39350 + }, + { + "epoch": 1.5197497972894707, + "grad_norm": 1.0977380275726318, + "learning_rate": 9.868592094932881e-05, + "loss": 0.229, + "step": 39360 + }, + { + "epoch": 1.5201359125834975, + "grad_norm": 0.5620018839836121, + "learning_rate": 9.866017992972702e-05, + "loss": 0.2642, + "step": 39370 + }, + { + "epoch": 1.5205220278775242, + "grad_norm": 0.36996108293533325, + "learning_rate": 9.863443891012524e-05, + "loss": 0.2314, + "step": 39380 + }, + { + "epoch": 1.520908143171551, + "grad_norm": 0.9804339408874512, + "learning_rate": 9.860869789052345e-05, + "loss": 0.2399, + "step": 39390 + }, + { + "epoch": 1.521294258465578, + "grad_norm": 0.4157778024673462, + "learning_rate": 9.858295687092166e-05, + "loss": 0.3006, + "step": 39400 + }, + { + "epoch": 1.5216803737596045, + "grad_norm": 0.5548539161682129, + "learning_rate": 9.855721585131988e-05, + "loss": 0.1816, + "step": 39410 + }, + { + "epoch": 1.5220664890536315, + "grad_norm": 0.9476989507675171, + "learning_rate": 9.853147483171809e-05, + "loss": 0.3943, + "step": 39420 + }, + { + "epoch": 1.522452604347658, + "grad_norm": 0.5183500647544861, + "learning_rate": 9.85057338121163e-05, + "loss": 0.1629, + "step": 39430 + }, + { + "epoch": 1.522838719641685, + "grad_norm": 1.4146567583084106, + "learning_rate": 9.847999279251452e-05, + "loss": 0.3829, + "step": 39440 + }, + { + "epoch": 1.5232248349357118, + "grad_norm": 2.4880552291870117, + "learning_rate": 9.845425177291273e-05, + "loss": 0.3052, + "step": 39450 + }, + { + "epoch": 1.5236109502297386, + "grad_norm": 0.43657195568084717, + "learning_rate": 9.842851075331094e-05, + "loss": 0.1691, + "step": 39460 + }, + { + "epoch": 1.5239970655237653, + "grad_norm": 0.13798825442790985, + "learning_rate": 9.840276973370916e-05, + "loss": 0.3217, + "step": 39470 + }, + { + "epoch": 1.5243831808177921, + "grad_norm": 0.8712138533592224, + "learning_rate": 9.837702871410737e-05, + "loss": 0.2702, + "step": 39480 + }, + { + "epoch": 1.524769296111819, + "grad_norm": 1.155957579612732, + "learning_rate": 9.83512876945056e-05, + "loss": 0.2693, + "step": 39490 + }, + { + "epoch": 1.5251554114058457, + "grad_norm": 1.194615364074707, + "learning_rate": 9.83255466749038e-05, + "loss": 0.1541, + "step": 39500 + } + ], + "logging_steps": 10, + "max_steps": 77697, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.405383667712e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/training_args.bin differ